Automated Syllabus of Machine Learning Papers

Built by Rex W. Douglass @RexDouglass ; Github ; LinkedIn

Papers curated by hand, summaries and taxonomy written by LLMs.

Submit paper to add for review

Introduction

History Of Machine Learning

  • Be aware of the unique challenges posed by machine learning systems, particularly in terms of technical debt, and adopt strategies to manage and minimize this debt throughout the entire lifecycle of the project. (Ananthanarayanan et al. 2013)

  • Aim to create a universally applicable and formalized definition of intelligence that does not rely on specific sets of senses, environments, or hardware, and that can effectively serve as a test for evaluating the intelligence of diverse systems. (NA?)

  • Carefully select appropriate machine learning algorithms based on your specific needs, and always validate your models using a separate hold-out dataset to avoid overfitting. (NA?)

  • Explore the potential of integrating quantum mechanics principles into machine learning algorithms to potentially achieve significant improvements in computational efficiency and accuracy. (NA?)

  • Aim to create computational models that demonstrate improvement over time, revealing underlying principles of learning applicable across various domains and representations. (NA?)

Applications Of Machine Learning

  • Utilise the geomstats Python package for performing computations on Riemannian manifolds, as it provides efficient and extensively unit-tested implementations of these manifolds, along with useful Riemannian metrics and associated exponential and logarithmic maps. (Miolane et al. 2018)

Basic Principles And Methods In Machine Learning

  • Carefully consider the potential for unintended feature leakage in collaborative machine learning systems, as this can lead to privacy violations such as membership inference and property inference attacks. (Carlini et al. 2018)

  • Utilise a layered architecture approach when creating a low-latency online prediction serving system, whereby the model abstraction layer handles the heterogeneous nature of existing machine learning frameworks and models, and the model selection layer dynamically selects and combines predictions across competing models to enhance accuracy and robustness. (Alekh Agarwal et al. 2016)

  • Utilize Bayesian teaching, a methodology that selects a small subset of data to effectively communicate the inferences of a machine learning model, thereby enhancing the explainability of these models. (Kelvin Xu et al. 2015)

  • Carefully evaluate various metric learning algorithms based on your unique properties, such as learning paradigm, form of metric, scalability, optimality of the solution, and dimensionality reduction, before selecting the most suitable method for your specific problem. (Bellet, Habrard, and Sebban 2013)

  • Use a combination of psychological and mathematical approaches to develop a robust learning method that can handle noisy data and changing concepts over time, as demonstrated by the STAGGER program. (NA?)

  • Adopt a two-step approach to process mining, involving the generation of a transition system as an intermediate representation, followed by its transformation into a Petri net using region theory. This enables better control over the degree of generalisation during the creation of the transition system, thereby helping to strike a balance between overfitting’ and ‘underfitting’. (NA?)

  • Use the (_{p})-norm multiple kernel learning methodology for improved efficiency and accuracy when dealing with multiple kernel learning problems, as demonstrated through empirical applications in bioinformatics and computer vision. (NA?)

  • Use a non-parametric resampling approach to determine the optimal split for your dataset, rather than relying on common rules-of-thumb like allocating 2/3rd of cases for training, especially if they have a smaller dataset size (n) and need higher classification accuracy. (NA?)

  • Investigate the optimal balance between prediction accuracy and explainability in AI systems, considering the varying needs of different stakeholders and application areas, to foster trustworthiness, fairness, and informed decision-making. (NA?)

  • Carefully consider the appropriate machine learning algorithm to use based on the nature of the available data and the desired outcome, as different algorithms have varying strengths and limitations. (NA?)

  • Utilise an online optimization algorithm for dictionary learning, specifically designed for sparse coding, which scales up gracefully to large datasets with millions of training samples, resulting in faster performance and better dictionaries than traditional batch algorithms. (NA?)

Supervised Learning Algorithms

  • Utilize the proposed importance sampling algorithm for nonparametric models given exchangeable binary response data, as it allows for efficient calculation of the permanent of a specific class of (0,1)-matrices in polynomial time, enabling accurate estimation of the marginal likelihood and subsequent posterior inference. (Christensen 2024)

  • Utilize the fused extended two-way fixed effects’ (FETWFE) estimator when dealing with difference-in-differences under staggered adoption scenarios. This estimator, based on machine learning techniques, automatically selects the necessary restrictions to balance bias reduction and efficiency improvement, thereby enhancing the accuracy of the analysis.’ (Faletto 2023)

  • Use the Root Causal Inference with Negative Binomials (RCI-NB) algorithm to account for measurement errors and counts in scRNA-seq data, allowing them to identify patient-specific root causes of diseases without requiring prior knowledge of the underlying structural equations or counterfactual distributions. (E. V. Strobl 2023)

  • Utilise a semiparametric functional factor model (SFFM) to bridge the gap between parametric and nonparametric functional data models. This model combines a parametric template with a nonparametric and infinite-dimensional basis expansion for the functions, allowing for greater flexibility and distinctness between the parametric and nonparametric components. (Kowal and Canale 2023)

  • Utilize a fully Bayesian Improved Surname Geocoding (fBISG) methodology along with name supplements to enhance the accuracy of race imputation, particularly for racial minorities, by addressing census data problems such as zero counts and missing surnames. (Rosenman, Olivella, and Imai 2022)

  • Utilise a Bayesian approach for data-driven discovery of non-linear spatio-temporal dynamic equations, which allows for the accommodation of measurement noise and missing data, and accounts for parameter uncertainty. (North, Wikle, and Schliep 2022)

  • Use mBART, a constrained version of BART, to improve the interpretability, predictive accuracy, and reduce post-data uncertainty in regression models involving monotone relationships between variables. (Chipman et al. 2022)

  • Utilize Bayesian methods for regression and classification problems, specifically the Relevance Vector Machine (RVM) model, which overcomes several limitations of the commonly used Support Vector Machine (SVM) while maintaining its desirable sparsity property. (Fradi et al. 2022)

  • Utilize non-parametric regression-based methods to estimate heterogeneous treatment effects in observational data, taking care to address issues such as selection bias, partial overlap, and unconfoundedness. (A. Caron, Baio, and Manolopoulou 2022)

  • Utilise the VadaBoost algorithm, which is based on sample variance penalisation, instead of traditional empirical risk minimisation techniques like AdaBoost. This is due to the fact that VadaBoost provides a balance between the sample mean and the sample variance of the exponential loss, leading to improved performance and handling of various types of weak learners. (“Planning for Mobile Manipulation” 2021)

  • Utilize the mixgb’ framework for multiple imputation, which combines XGBoost, subsampling, and predictive mean matching to effectively handle large datasets with complex data structures, reducing bias and enhancing imputation quality.’ (Yongshi Deng and Lumley 2021)

  • Utilise a three-stage estimation process for efficient nonparametric estimation of generalized panel data transformation models with fixed effects. (Liang Jiang et al. 2021)

  • Utilize a fast rejection sampling technique for the Conway-Maxwell-Poisson distribution to improve computational efficiency and reduce central processing unit (CPU) time in performing inference for COM-Poisson regression models. (Benson and Friel 2021)

  • Adopt a time-adaptive approach to exploring, weighting, combining, and selecting models that differ in terms of predictive variables included, allowing for changes in the sets of favored models over time, and guiding this adaptivity by the specific forecasting goals. (I. Lavine, Lindon, and West 2021)

  • Utilize the Partial Fourier Transform (PFT) algorithm instead of the traditional Fast Fourier Transform (FFT) for more efficient and accurate computation of partial Fourier coefficients, particularly when dealing with large input lengths or numerous FFT operations. (Y. Park, Jang, and Kang 2021)

  • Develop a two-stage approach for recommending the appropriate package type for e-commerce shipments, taking into account the trade-offs between shipping and damage costs, and utilizing a scalable, computationally efficient linear time algorithm. (Gurumoorthy, Sanyal, and Chaoji 2020)

  • Aim to generate prediction intervals that have a user-specified coverage level across all regions of feature-space, a property called “conditional coverage”, by modifying the loss function to promote independence between the size of the intervals and the indicator of a miscoverage event. (Yichen Jia and Jeong 2020)

  • Carefully evaluate the assumptions, philosophies, and goals of both traditional regression methods and newer pure prediction algorithms when selecting the optimal approach for your specific research context. (Efron 2020)

  • Consider utilising a unified boosting algorithm across multiple classifier graphs, allowing for the development of simple, efficient, and highly accurate boosting algorithms tailored to specific types of classifiers. (Valdes et al. 2020)

  • Use temporal residual based metrics to evaluate cross-validation efforts in binary-time-series-cross-section data, rather than traditional classification metrics, to avoid underestimation of model performance. (Çiflikli et al. 2019)

  • Simplify traditional two-stage methods for non-linear instrumental variable (IV) regression by using a dual formulation, enabling them to avoid the first-stage regression which can be a bottleneck in real-world applications. (Muandet et al. 2019)

  • Consider using kernel instrumental variable regression (KIV) as a nonparametric generalization of traditional two-stage least squares (2SLS) algorithms for estimating causal effects in observational data, particularly when the underlying relationships are likely to be nonlinear. (R. Singh, Sahani, and Gretton 2019)

  • Utilise the conformalized quantile regression’ (CQR) method when seeking to generate accurate prediction intervals in regression modelling. This method combines the benefits of conformal prediction - which provides a nonasymptotic, distribution-free coverage guarantee - with the efficiency of quantile regression, allowing for the generation of prediction intervals that are adaptive to heteroscedasticity. (Vovk et al. 2019)

  • Consider using Thresholded EEBoost (ThrEEBoost) for variable selection in messy high-dimensional datasets, as it enables exploration of diverse variable selection paths and potentially leads to models with lower prediction error. (Speiser et al. 2019)

  • Consider developing a typology of performance metrics to enhance understanding of your structure and properties, thereby improving the selection process in machine learning regression, forecasting, and prognostics. (Botchkarev 2019)

  • Focus on developing a hierarchical indexing structure based on Vector and Bilayer Line Quantization (VBLQ) to improve the efficiency and accuracy of approximate nearest neighbor (ANN) searches on GPUs. (Wei Chen et al. 2019)

  • Consider reformulating the related searches problem into an extreme classification task, utilize the Slice algorithm for extreme multi-label learning with low-dimensional dense features, and evaluate its performance against existing techniques to demonstrate its potential benefits in increasing trigger coverage, suggestion density, and recommendation accuracy. (H. Jain et al. 2019)

  • Carefully choose the appropriate gradient boosting decision tree (GBDT) algorithm depending on the specific learning task and dataset characteristics, considering factors such as GPU acceleration capabilities, hyper-parameter optimization strategies, and overall generalization performance. (Anghel et al. 2018)

  • Focus on hypothesis 3, which states that identifying a robust classifier from limited training data is information theoretically possible but computationally intractable, as it provides strong evidence for the possibility of robust classification tasks that are information theoretically easy but computationally intractable under a powerful model of computation (statistical query model). (Bubeck, Price, and Razenshteyn 2018)

  • Consider using polynomial regression models as an alternative to neural networks, as they offer comparable accuracy and avoid common pitfalls associated with neural network models, such as hyperparameter tuning and convergence issues. (Xi Cheng et al. 2018)

  • Utilize SHAP values for tree ensemble feature attribution due to its consistency, local accuracy, and ability to handle missingness, providing a strict theoretical improvement over existing methods like the Saabas method. (Lundberg, Erion, and Lee 2018)

  • Consider using lossless compression methods for large tree-based ensemble models, specifically random forests, to address the issue of increased storage requirements caused by growing dataset sizes and complexities. (Painsky and Rosset 2018)

  • Prioritize developing safe semi-supervised learning techniques that ensure the generalization performance is never statistically significantly worse than methods using only labeled data, especially considering factors such as data quality, model uncertainty, and measure diversity. (Q. Yao et al. 2018)

  • Conduct average-case analyses of specific algorithms, taking into consideration the target concept, number of irrelevant attributes, and class and attribute frequencies, to obtain accurate predictions about the behavior of induction algorithms and validate your analyses through experimentation. (J. Luo, Meng, and Cai 2018)

  • Apply robust optimization principles to model the noise arising in online advertising signals as bounded box-type interval uncertainty sets, and develop robust factorization machine (RFM) and robust field-aware factorization machine (RFFM) algorithms as robust minimax formulations for FM and FFM respectively. (Punjabi and Bhatt 2018)

  • Use a gradient boosting machine for function approximation, which is a powerful tool for optimizing numerical problems in function space, particularly useful for handling complex datasets and producing accurate predictions. (Martínez-Velasco, Martínez-Villaseñor, and Miralles-Pechuán 2018)

  • Prioritise privacy-aware feature selection and composition, utilising minimum and maximum based composition among raw features, and employing a hybrid tree ensemble model selection approach to achieve optimal performance. (S. Ji et al. 2018)

  • Use Selective Gradient Boosting (SelGB) to effectively rank items by focusing on the most informative negative examples during the learning process, thereby improving the overall performance of your model. (Lucchese et al. 2018)

  • Utilize classifier systems, which are massively parallel, message-passing, rule-based systems that learn through credit assignment (using the bucket brigade algorithm) and rule discovery (via the genetic algorithm), to address challenges posed by perpetually novel events, noisy or irrelevant data, continuous real-time requirements for action, implicitly or inexactly defined goals, and sparse payoffs or reinforcement obtained only through long action sequences. (“Encyclopedia of Machine Learning and Data Mining” 2017)

  • Carefully consider the goals of your analysis and choose appropriate methods accordingly, balancing the tradeoff between providing valid confidence intervals and achieving out-of-sample predictive power. (Arjovsky and Bottou 2017)

  • Utilise the ggRandomForests package when working with Random Forest Survival Models to enhance visualisation and interpretation of the model, thereby improving its applicability and usefulness. (Ehrlinger 2016)

  • Implement the ordering principle’ to solve issues related to target leakage and prediction shift in gradient boosting algorithms, resulting in improved performance through the use of ‘ordered boosting’ and a novel algorithm for processing categorical features.’ (Ferov and Modrý 2016)

  • Consider using a Bayesian probabilistic framework for learning in general models of the form (1), which offers good generalization performance and produces exceedingly sparse predictors containing relatively few non-zero parameters. (Senekane and Taele 2016)

  • Aim to create a general framework for variance reduction in online experiments using advanced machine learning techniques, such as gradient boosted decision trees, to improve the accuracy and efficiency of A/B testing in internet companies. (Poyarkov et al. 2016)

  • Utilize appropriate evaluation metrics tailored to the specific needs of imbalanced datasets, rather than relying solely on standard metrics like accuracy or mean squared error, which may not accurately reflect the performance of models in these scenarios. (Branco, Torgo, and Ribeiro 2015)

  • Focus on developing accurate models for intermolecular forces and combine them with the GDML model to enable predictive simulations of condensed molecular systems. (Hirn, Poilvert, and Mallat 2015)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Leek and Peng 2015)

  • Utilise the mldr package in R to effectively explore, analyse, and manipulate multilabel datasets, enabling accurate prediction and classification. (Charte and Charte 2015)

  • Consider using QuickScore, a novel algorithm designed for efficiently ranking documents through the use of additive ensembles of regression trees, as it offers significant improvements in computational speed without compromising on accuracy. (Lucchese et al. 2015)

  • Utilize a robust random cut forest (RRCF) data structure for efficient anomaly detection in dynamic data streams, as it effectively preserves distances and enables accurate identification of anomalous points based on your impact on the overall dataset. (Lavin and Ahmad 2015)

  • Consider the potential impact of task-induced bias when conducting class incremental learning studies, and explore ways to minimize this bias through causal interventions and debias modules. (G. Hinton, Vinyals, and Dean 2015)

  • Consider using the newly proposed family of one-factor distributions for high-dimensional binary data, which offers an explicit probability for each event, easy model interpretation, and efficient parameter estimation via the inference margin procedure and expectation-maximization algorithm. (Marbac and Sedki 2015)

  • Consider using the Fastfood algorithm to efficiently approximate kernel expansions in loglinear time, providing significant speedups compared to traditional methods without sacrificing accuracy. (Quoc Viet Le, Sarlos, and Smola 2014)

  • Adopt Kernel Regularized Least Squares (KRLS) for social science modeling and inference problems, as it combines the flexibility of machine learning techniques with the interpretability of traditional statistical models, reducing misspecification bias and enabling robust conclusions. (Hainmueller and Hazlett 2014)

  • Use a scalable machine learning framework based on maximum entropy (logistic regression) to address the challenge of predicting user response in display advertising, while incorporating feature hashing to manage the high dimensionality of the data. (Chapelle, Manavoglu, and Rosales 2014)

  • Carefully consider the trade-off between the accuracy and cost of oracle measurements when developing a bandit strategy for optimizing demographic targeting in digital advertising. (M. H. Williams et al. 2014)

  • Consider using the Laplace distribution instead of the traditional Gaussian distribution when dealing with sparse data in factorization machines for click-through rate prediction tasks. (Baqapuri and Trofimov 2014)

  • Focus on developing methods that balance bias and variance in statistical models, using techniques like distributionally robust optimization and Owens empirical likelihood to create convex surrogates for variance, leading to more accurate and efficient modeling.’ (Bertsimas, Gupta, and Kallus 2014b)

  • Ensure comparability among different approaches by standardising datasets, protocols, and computational budgets, and that they should prioritise optimisation methods that balance running time and accuracy in multi-codebook quantization tasks. (Bezanson et al. 2014)

  • Consider extending the Local Sensitivity Hashing (LSH) framework to include asymmetric hashing schemes, allowing for efficient sublinear hashing algorithms for Maximum Inner Product Search (MIPS) problems. (Shrivastava and Li 2014)

  • Use distance-induced kernels to resolve the issue of nonintegrability of weight functions in order to establish the link between RKHS-based dependence measures and the distance covariance. (Sejdinovic et al. 2013)

  • Be aware that large-sample learning of Bayesian networks is NP-hard, meaning that identifying high-scoring structures is computationally difficult even when using a consistent scoring criterion and having access to an independence oracle, inference oracle, or information oracle. (Chickering, Heckerman, and Meek 2013)

  • Adopt a 5-fold cross validation strategy when using the LETOR 4.0 datasets, ensuring they divide your data into separate training, validation, and testing sets within each fold. (Tao Qin and Liu 2013)

  • Utilize the Sparse Least Trimmed Squares (Sparse LTS) estimator when dealing with high dimensional datasets containing outliers, as it provides both robustness against outliers and sparsity in model estimates, thus enhancing interpretability and prediction accuracy. (Alfons, Croux, and Gelper 2013)

  • Utilize Individual Conditional Expectation (ICE) plots rather than traditional Partial Dependence Plots (PDPs) to effectively visualize the impact of specific features on the predicted outcome in supervised learning algorithms, especially when dealing with significant interaction effects. (Goldstein et al. 2013)

  • Consider implementing a collaborative boosting framework for activity classification in microblogs, which involves maintaining separate classifiers for each user and allowing collaboration between those classifiers based on shared training instances and dynamically changing labeling decisions. (Yangqiu Song et al. 2013)

  • Utilise the AdaBoost.MH algorithm with Hamming Trees for multi-class classification tasks due to its superior performance compared to other known implementations of AdaBoost.MH and its ability to perform on par with the best existing multiclass boosting algorithm AOSOLogitBoost and Support Vector Machines (SVMs). (Kégl 2013)

  • Focus on developing algorithms for learning kernels based on the concept of “centered alignment,” which measures the similarity between kernels or kernel matrices and has been shown to correlate strongly with improved performance in classification and regression tasks. (Cortes, Mohri, and Rostamizadeh 2012)

  • Utilise the novel techniques of “Gradient-based One-Side Sampling” (GOSS) and “Exclusive Feature Bundling” (EFB) to significantly enhance the efficiency and scalability of Gradient Boosting Decision Trees (GBDT) in scenarios involving high dimensionality and large data sizes. (Ping Li 2012)

  • Consider using a semi-parametric Bayesian framework for simultaneous analysis of linear quantile regression models, as it allows for a more comprehensive understanding of the relationships between variables while accounting for the monotonicity constraint inherent in quantile regression. (Tokdar and Kadane 2012)

  • Utilise a recursive partitioning algorithm to create a regression tree model that effectively analyses establishment nonresponse in surveys. This model provides mutually exclusive cells based on establishment characteristics with homogenous response propensities, allowing for easy interpretation of the associations between these characteristics and an establishment’s propensity to respond. Furthermore, the model can be tested against disjoint sets of establishment data to ensure its accuracy. (Phipps and Toth 2012)

  • Consider using Venn-Abers predictors for calibration in decision trees, as it provides a highly competitive approach that significantly outperforms Platt scaling, Isotonic regression, and no calibration across numerous performance metrics, except for AUC. (Vovk and Petej 2012)

  • Consider combining boosting algorithms with error-correcting output codes (ECOC) to improve the performance of multiclass learning problems, while maintaining the simplicity of binary classification tasks. (Mukherjee and Schapire 2011)

  • Employ a joint statistical model for multiple climate model errors that accounts for the spatial dependence of individual models as well as cross-covariance across different climate models, offering a nonseparable cross-covariance structure. (Sang, Jun, and Huang 2011)

  • Utilise a nonparametric modelling approach for degradation processes, especially when dealing with incomplete or sparsely observed degradation signals. (R. R. Zhou, Serban, and Gebraeel 2011)

  • Utilize a bivariate metric that combines both the variability of the estimate and the accuracy of classifying positive and negative users when developing multi-touch attribution models for digital advertising. (X. Shao and Li 2011)

  • Apply Structural Risk Minimization (SRM) principles to break down your hypothesis set into subsets of varying complexities and choose a base learner from a subset that offers the best trade-off between proximity to the functional gradient and complexity. (Grubb and Bagnell 2011)

  • Use a combination of statistical analysis and machine learning methods, specifically support vector machines (SVMs), to identify the most relevant clinical features for accurately predicting the presence of a STAT3 mutation in patients with Hyperimmunoglobulin E Syndrome (HIES). (Woellner et al. 2010)

  • Also pay attention to various parameters in the titan() function, such as the minimum number of observations on either side of a change point, the number of random permutations, and the number of bootstrap replications, to achieve optimal performance and accuracy in your analysis (M. E. Baker and King 2010)

  • Consider using the Searn meta-algorithm for structured prediction tasks, which involves treating these tasks as search problems and iteratively improving upon an initial classifier based on its performance on a series of cost-sensitive examples. (Daumé, Langford, and Marcu 2009)

  • Use a novel weak learnability formulation (lemma 8) that is more suitable for analyzing LogitBoost compared to previous formulations. (Ping Li 2009)

  • Utilize the Bolasso technique, which involves running the Lasso for several bootstrapped replications of a given sample and intersecting the supports of the Lasso bootstrap estimates, leading to consistent model selection without requiring the consistency condition needed by the standard Lasso. (F. Bach 2008)

  • Utilize a Bayesian “sum-of-trees” model called BART, which combines multiple weak learners through an iterative backfitting MCMC algorithm, allowing for accurate prediction and comprehensive uncertainty estimation. (Chipman, George, and McCulloch 2007)

  • Utilize the TMVA toolkit within the ROOT framework to effectively apply multivariate classification and regression techniques in high-energy physics, thereby maximizing the extraction of useful information from increasingly complex datasets. (Hoecker et al. 2007)

  • Consider implementing the Look-ahead Linear Regression Trees (LLRT) algorithm, which enables a near-exhaustive evaluation of all possible splits in a node, leading to improved predictive accuracy for problems with strong mutual dependencies between attributes. (Vogel, Asparouhov, and Scheffer 2007)

  • Conduct large-scale empirical evaluations of various supervised learning algorithms using multiple performance criteria to identify the strengths and weaknesses of each approach and inform future applications. (Caruana and Niculescu-Mizil 2006)

  • Utilize a Bayesian approach to fitting general design generalized linear mixed models (GLMMs) using Markov Chain Monte Carlo (MCMC) techniques, as it enables better handling of complex random effects structures and accounts for uncertainty in variance components. (Y. Zhao et al. 2006)

  • Seek a balanced approach between maximizing the error-correcting ability of the coding matrix and minimizing the difficulty of the binary problems generated for the base learner, as focusing solely on either aspect could lead to suboptimal performance in multiclass classification tasks. (Ling Li 2006)

  • Consider developing cost-sensitive boosting algorithms to improve the classification performance of imbalanced data involving multiple classes, particularly when the cost matrix is unknown, by utilizing genetic algorithms to search for the optimum cost setup of each class. (Yanmin Sun, Kamel, and Wang 2006)

  • Carefully choose your instrumental variables, data prefiltering, and extended IV criterion norm to optimize the performance of your closed-loop system identification studies. (Gilson and Hof 2005)

  • Consider using Iterated Bagging (IB) instead of Stochastic Gradient Boosting (SGB) for bias-variance reduction in regression problems, as IB consistently outperforms SGB across various datasets and scenarios. (“Machine Learning: ECML 2005” 2005)

  • Consider adopting a Bayesian approach to P-splines for modelling nonlinear smooth effects of covariates within the generalized additive and varying coefficient models framework, as it allows for simultaneous estimation of smooth functions and smoothing parameters, and can be easily extended to more complex formulations. (Lang and Brezger 2004)

  • Utilise the concept of Levy trees, which are continuous analogues of discrete Galton-Watson trees, to better understand the probabilistic properties of complex systems. (Duquesne and Gall 2004)

  • Focus on understanding the properties of the marginal likelihood function in order to optimize the performance of sparse Bayesian learning methods. (Faul and Tipping 2002)

  • Compare various discrimination methods for the classification of tumors based on gene expression profiles, including traditional techniques like nearest-neighbor and linear discriminant analysis, as well as newer machine learning approaches like bagging and boosting, across multiple datasets to determine the best approach for accurate and reliable classification. (Dudoit, Fridlyand, and Speed 2002)

  • Utilize the Bayesian Committee Machine (BCM) technique for combining multiple estimators trained on separate datasets, particularly in situations involving kernel-based regression systems and large data sets. (Tresp 2000)

  • Understand boosting as a technique for fitting an additive model, rather than focusing solely on improving the performance of individual classifiers through a weighted majority vote or committee. (J. Friedman, Hastie, and Tibshirani 2000)

  • Consider applying lazy learning techniques to Bayesian tree induction, specifically through the development of a lazy Bayesian rule learning algorithm (Lbr), which can lead to reduced error rates compared to traditional methods like naive Bayesian classifiers, C4.5, Bayesian tree learning algorithms, and even selective naive Bayesian classifiers. (Zijian Zheng and Webb 2000)

  • Develop performance bounds for model selection criteria using recent theory for sieves, focusing on the problem of estimating the unknown density or regression function, and aiming for simultaneous minimax rate optimality across multiple classes of smoothness depending on the chosen list of models. (Barron, Birgé, and Massart 1999)

  • Consider employing a two-step estimation procedure when dealing with varying coefficient models, especially when the coefficient functions exhibit differing levels of smoothness. This approach offers improved accuracy and reliability compared to traditional one-step approaches, while remaining relatively insensitive to the choice of initial bandwidth. (J. Fan and Zhang 1999)

  • Utilise a Winnow-based algorithm for context-sensitive spelling correction, as it demonstrates superior performance over traditional Bayesian methods, especially when handling larger feature sets. (Golding and Roth 1998)

  • Consider using a Bayesian approach to curve fitting, specifically through the use of piecewise polynomials with an unknown number of knots at unknown locations, allowing for the estimation of a wide range of curve shapes while avoiding issues related to overparameterization and underparameterization. (Denison, Mallick, and Smith 1998)

  • Focus on improving the margin of your models, i.e., the difference between the weight assigned to the correct label and the maximum weight assigned to any incorrect label, as doing so leads to a reduced generalization error. (P. Bartlett et al. 1998)

  • Utilize a broad spectrum of classifiers across various domains and implement rigorous parameter tuning to ensure fair and comprehensive evaluations of classifier performances. (Aha, Kibler, and Albert 1991)

  • Aim to develop algorithms that balance the need for accurate classification with the desire for simple, comprehensible rules, while maintaining efficiency in rule generation, particularly when working with noisy data. (P. Clark and Niblett 1989)

  • Aim to develop algorithms that balance the need for accurate classification with the desire for simple, comprehensible rules, while maintaining efficiency in rule generation, particularly when working with noisy data. (P. Clark and Niblett 1989)

  • Use the observable window ({e}) instead of the unobservable optimal window (h{0}) when comparing different data-driven approaches to determine window size in nonparametric density estimation, because ({e}) performs just as well as (h{0}) to both first and second order. (Hall and Marron 1987)

  • Use local weighted polynomial regression to estimate parameters in your models, as it provides an asymptotically optimal estimator under minimal assumptions about the underlying data. (Kliemann 1987)

  • Use a local linear smoother with variable bandwidth to improve your estimates accuracy and flexibility in handling complex shapes of regression functions.’ (Kliemann 1987)

  • Utilize local polynomial fitting directly as a weighted least squares estimator instead of an approximate kernel estimator to simplify the understanding of asymptotic behavior, especially in complex scenarios like multivariate x, higher polynomials, or derivative estimation. (Kliemann 1987)

  • Utilise a novel method for flexible regression modelling of high dimensional data, which uses an expansion in product spline basis functions. This method allows for automatic determination of the number of basis functions, product degree, and knot locations, providing greater power and flexibility to model relationships that are nearly additive or involve interactions in just a few variables. (Kliemann 1987)

  • Utilise the Alternating Conditional Expectations (ACE) algorithm to identify optimal transformations for your data, thereby improving the accuracy of your statistical inferences. (Breiman and Friedman 1985)

  • Utilize the Bayesian approach to modeling, specifically the dynamic generalized linear model (DGLM), because it offers advantages over traditional generalized linear models (GLMs) by allowing for sequential analysis, closed form updating and predictive distributions, and computational simplicity. (West, Harrison, and Migon 1985)

  • Carefully choose the appropriate statistical model and estimation strategy for your study, taking into consideration factors such as sample size, measurement errors, missing data, and potential confounding variables. (Haskell and Hanson 1981)

  • Utilize the Smoothed Cross-Validation (SCV) method for selecting the bandwidth of a kernel density estimator, as it offers superior performance compared to traditional Least Squares Cross-Validation (CV) due to its ability to reduce sample variability without sacrificing accuracy. (Strassen 1964)

  • Use Empirical Risk Minimization (ERM) classifiers to achieve optimal rates in statistical learning tasks, particularly when dealing with massive datasets, while being mindful of the margin parameter and the complexity of the class of possible sets. (Stevens 1946)

  • Understand boosting as a technique for fitting an additive model, rather than focusing solely on improving the performance of individual classifiers through a weighted majority vote or committee. (NA?)

  • Utilize local polynomial fitting directly as a weighted least squares estimator instead of an approximate kernel estimator to simplify the understanding of asymptotic behavior, especially in complex scenarios like multivariate x, higher polynomials, or derivative estimation. (NA?)

  • Utilise a novel method of flexible nonparametric regression modelling that uses product spline basis functions to represent the relationship between a response variable and multiple predictors. This method offers advantages over traditional approaches like recursive partitioning and additive modelling because it allows for greater flexibility and power in modelling relationships that are nearly additive or involve interactions among just a few variables. Additionally, the model can be expressed in a way that separates the additive components from the multi-variable interactions (NA?)

  • Consider using quantile regression techniques when estimating a specific quantile of a dependent variable, instead of focusing solely on the conditional mean, as it provides valuable insights into the distribution of the random variable. (NA?)

  • Utilize the Nested Generalized Exemplar (NGE) learning method, which involves storing objects in Euclidean n-space as hyperrectangles that can be nested inside one another to arbitrary depth, allowing for efficient storage and retrieval of information while preserving the original structure of the data. (NA?)

  • Carefully consider the choice between decision bound and exemplar models when analyzing categorization data, as the former may offer superior explanatory power in certain situations. (NA?)

  • Utilize the RELIEF algorithm, specifically its extension RELIEF-F, for estimating attributes in multi-class problems, as it demonstrates superior performance over other methods in dealing with noisy, incomplete, and multi-class datasets. (NA?)

  • Carefully choose appropriate machine learning paradigms based on the specific requirements of your problem, considering aspects such as representation, performance methods, and learning algorithms. (NA?)

  • Consider using decision tables as a hypothesis space for supervised learning algorithms, particularly when dealing with discrete features, as they can often outperform more complex algorithms like C4.5 while being easier to interpret. (NA?)

  • Utilise the MEME algorithm, which expands upon the traditional expectation maximisation (EM) algorithm, to identify multiple motifs within unaligned biopolymer sequences. This is achieved through the use of subsequences that actually occur in the biopolymer sequences as starting points for the EM algorithm, removing the assumption that each sequence contains exactly one occurrence of the shared motif, and probabilistically erasing shared motifs after they are found. (NA?)

  • Consider using entropy as a distance measure in your studies, as it offers a unified approach to dealing with various challenges such as handling symbolic attributes, real valued attributes, and missing values. (NA?)

  • Consider using the Recurrence Surface Approximation (RSA) technique when dealing with censored data in medical contexts, as it provides a robust and effective way to predict Time to Recur (TTR) based on a linear combination of input features. (NA?)

  • Carefully choose appropriate performance metrics when dealing with imbalanced datasets, as traditional methods like accuracy may lead to misleading conclusions. (NA?)

  • Focus on the relationship between boosting and support vector machines, recognizing that both can be seen as methods for regularized optimization in high-dimensional predictor space, with boosting providing an approximate path to maximum margin classifiers. (NA?)

  • Consider utilizing a unifying framework for solving multiclass categorization problems by reducing them to multiple binary problems, which can then be addressed using a margin-based binary learning algorithm. (NA?)

  • Consider implementing an online SVM algorithm, specifically LASVM, due to its efficiency in handling large datasets, achieving competitive misclassification rates after just one pass through the training examples, and requiring less memory compared to state-of-the-art SVM solvers. (NA?)

  • Analyze learning curves to determine the optimal choice between logistic regression and tree induction for a given dataset, as the preference for one method over the other depends on factors like training set size and separability of signal from noise. (NA?)

  • Focus on understanding the underlying principles of learning theory, particularly the role of the regression function and the importance of minimizing the error in order to accurately predict outputs based on inputs. (NA?)

  • Utilize ultraconservative algorithms for multiclass problems, which involve updating only the prototypes attaining similarity-scores higher than the score of the correct labels prototype, leading to improved performance and efficiency.’ (NA?)

  • Focus on finding the optimal regularization parameter (γ) to minimize the error between the approximated function (f_γ,z) and the true regression function (f_ρ) when using the proposed approach in learning theory. (NA?)

  • Consider incorporating fuzzy membership into your support vector machine models to account for varying levels of importance among input points, thereby enhancing model accuracy and robustness against noise and outliers. (NA?)

  • Consider applying lazy learning techniques to Bayesian tree induction, specifically through the development of the lazy Bayesian rule learning algorithm (LBR), which demonstrates improved performance over traditional methods like naive Bayesian classifiers, C4.5, Bayesian tree learning algorithms, and others across various natural domains. (NA?)

  • Adopt a framework for sparse Gaussian processes (GP) methods that uses forward selection with criteria based on information-theoretic principles, allowing for efficient learning of d-sparse predictors and effective training under strict time and memory constraints. (NA?)

  • Consider adopting sparse Bayesian learning (SBL) for basis selection tasks due to its ability to prevent structural errors and potentially possess fewer local minima than existing alternatives, leading to improved performance. (NA?)

  • Utilise Maximum Entropy Discrimination (MED) to develop Support Vector Machines (SVMs) that can perform feature selection and kernel selection tasks simultaneously, thereby enhancing the efficiency and accuracy of the SVMs. (NA?)

  • Consider using L1-based regularization instead of L2-based regularization for logistic regression when dealing with many features, as it leads to improved performance and reduced sample complexity. (NA?)

  • Utilise Gaussian Processes in Machine Learning due to your ability to provide a flexible, non-parametric modelling approach that enables accurate prediction and efficient handling of large datasets. (NA?)

  • Consider using sparse multinomial logistic regression (SMLR) for accurate and efficient classification tasks, especially when dealing with large datasets in high-dimensional feature spaces. (NA?)

  • Carefully evaluate the reliability and validity of your measuring procedures when conducting comparative studies of software prediction models, as the current commonly used measuring procedure has been found to be unreliable, potentially contributing to the lack of convergence in the field. (NA?)

  • Consider combining the advantages of both the Michigan and Pittsburgh approaches in fuzzy genetics-based machine learning (FGBML) algorithms to improve the efficiency and accuracy of finding fuzzy rule-based systems for pattern classification problems. (NA?)

  • Consider combining tree induction and logistic regression methods to create “logistic model trees” (LMT) for classification tasks, as this approach can provide more accurate and interpretable classifiers compared to traditional methods. (NA?)

  • Extend learning theory beyond scalar-valued functions to include vector-valued functions, using reproducing kernel Hilbert spaces and minimal norm interpolation techniques, in order to improve performance in various applications. (NA?)

  • Utilize the proposed two novel support vector approaches for ordinal regression, which optimize multiple thresholds to define parallel discriminant hyperplanes for the ordinal scales, ensuring proper ordering of thresholds at the optimal solution. (NA?)

  • Carefully consider the choice of loss function and basis functions in your boosting algorithms, as they significantly impact the performance and convergence properties of the model. (NA?)

  • Address the challenge of imbalanced datasets in medical diagnostics by employing prototype-based resampling or asymmetrical margin support vector machines to optimize model performance. (NA?)

  • Prioritise classifier performance over codeword separation when designing error correcting output codes (ECOC) matrices, leading to higher discriminatory power and reduced need for classifiers. (NA?)

  • Utilize computer-based models to understand complex adaptive systems (CAS), due to the limitations of traditional mathematical tools such as partial differential equations (PDEs) and statistical techniques in accurately capturing the nonlinear dynamics and continuous adaptation inherent in CAS. (NA?)

  • Utilise cost curves instead of ROC curves for visualising classifier performance due to your ability to provide instant answers to various critical experimental questions through visual inspection. (NA?)

  • Understand the importance of ROC graphs in organizing and visualizing classifier performance, particularly in situations involving skewed class distributions and unequal classification error costs, and avoid common misconceptions and pitfalls when using them in practice. (NA?)

  • Frame learning sequential, goal-directed behavior as a maximum margin structured prediction problem over a space of policies, allowing them to learn mappings from features to cost so an optimal policy in an MDP with these cost mimics the experts behavior. (NA?)

  • Utilise the Component Selection and Smoothing Operator (COSSO) method for model selection and estimation in SS-ANOVA, as it offers a robust and efficient approach compared to existing techniques like the LASSO and MARS procedures. (NA?)

  • Employ a convex optimization scheme to model shared characteristics as linear transformations of the input space, which can lead to significant improvements in the accuracy of multiclass linear classifiers. (NA?)

  • Conduct comprehensive experiments involving multiple datasets, various sampling techniques, and diverse learning algorithms to ensure robust, statistically valid, and reliable findings about the relative strengths and weaknesses of different techniques in handling imbalanced data. (NA?)

  • Utilize sparse optimization methods, specifically LASSO, to identify the underlying PDE governing a given dataset, promoting sparsity in the vector α and assuming that the underlying dynamics are governed by a few terms. (NA?)

  • Consider extending multiple kernel learning (MKL) to arbitrary norms, specifically (_{p})-norms with \(p >= 1\), to improve the robustness and generalizability of kernel mixtures. (NA?)

  • Adopt a probabilistic approach for supervised learning when faced with multiple annotators providing possibly noisy labels but no absolute gold standard, allowing for evaluation of different experts and estimation of the actual hidden labels. (NA?)

  • Consider using the ADASYN algorithm for handling imbalanced datasets, as it adaptively generates synthetic data for minority class samples based on your level of difficulty in learning, thereby reducing bias and focusing on hard-to-learn examples. (NA?)

  • Carefully consider and control for potential sources of bias in your experimental designs, particularly when comparing different classification algorithms like random forests and support vector machines. (NA?)

  • Consider implementing a special-purpose solver for the specific instance of semidefinite programming that arises in LMNN classification, allowing for scalability to larger datasets and improved performance. (NA?)

  • Carefully examine the properties of your loss functions, such as consistency, soundness, continuity, differentiability, and convexity, to ensure accurate and efficient learning to rank models. (NA?)

  • Carefully consider the choice of upscaling method for estimating carbon fluxes, as it significantly impacts the final results, and ensure adequate representation of the training dataset to minimize hidden extrapolations. (NA?)

  • Focus on developing a methodology that enables the creation of a quantizer that approximates a sufficient statistic for its attribute label, thereby allowing for accurate prediction of the attribute even when working with limited information. (NA?)

  • Carefully choose performance measures for classification tasks based on your invariance properties, as these properties directly impact the reliability and objectivity of the evaluation process. (NA?)

  • Focus on developing a diverse population of rules rather than searching for a single best-fit model when dealing with complex systems. (NA?)

  • Consider using binary relevance-based methods for multi-label classification tasks, as they offer significant benefits in terms of scalability and computational complexity, while still being able to effectively capture label correlations through techniques such as classifier chains. (NA?)

  • Consider using a combination of instance-based learning and logistic regression for multilabel classification tasks, as it allows for better representation of correlations between labels and provides an easily interpretable solution. (NA?)

  • Utilize the 1-slack formulation for structural SVMs, which replaces multiple cutting-plane models with a single one, resulting in a significant improvement in computational efficiency without sacrificing generalizability. (NA?)

  • Focus on developing a scalable, accurate, and efficient Bayesian click-through rate (CTR) prediction algorithm for sponsored search advertising, incorporating factors such as ad features, query features, and context features, while considering the unique challenges posed by the dynamic nature of the internet and the need for continuous updating and optimization. (NA?)

  • Utilize online learning algorithms for detecting malicious websites, as they can process large amounts of data more efficiently than batch methods and adapt to evolving patterns in malicious URLs over time. (NA?)

  • Utilize a probabilistic approach for supervised learning when dealing with multiple potentially noisy experts, rather than simply employing majority voting, because the former allows for better evaluation of individual experts and estimation of the actual hidden labels. (NA?)

  • Consider applying the Random Forests machine-learning algorithm to model complex and potentially non-linear relationships between oceanic properties and seafloor standing stocks, as it offers several advantages over traditional statistical methods. (NA?)

  • Consider combining multiple resampling techniques with cost-sensitive learning (CSL) to effectively address class imbalance issues in machine learning algorithms, leading to improved classifier performance and reduced misclassification costs. (NA?)

  • Carefully evaluate and specify the assumptions underlying your choice of multi-instance learning algorithms, as different problem domains may require distinct MI assumptions. (NA?)

  • Consider utilizing a unified decision forest framework for various machine learning, computer vision, and medical image analysis tasks, as it offers efficiency, versatility, and potential improvements over alternative approaches. (NA?)

  • Consider using the classifier chains method for multi-label classification tasks, as it effectively models label correlations while maintaining reasonable computational complexity. (NA?)

  • Conduct an exhaustive empirical study of OVO and OVA decompositions, focusing on various ways to combine the outputs of base classifiers, and analyze the behavior of these schemes with different base learners. (NA?)

  • Focus on developing a comprehensive framework for variable-star classification that includes proper feature creation and selection in the presence of noise and spurious data, fast and accurate classification, and improved classification through the use of taxonomy. (NA?)

  • Consider using unbiased classification tree algorithms like CRUISE, GUIDE, and QUEST, which utilize a two-step approach based on significance tests to split each node, ensuring that every X variable has an equal chance of being selected regardless of the number of distinct values it possesses. (NA?)

  • Differentiate between conditional and marginal label dependence in multi-label classification, as this distinction impacts the choice of appropriate loss functions and ultimately influences the predictive performance of the classifier. (NA?)

  • Utilise Receiver Operator Characteristics (ROC) curves instead of prediction accuracy for the assessment of biomarker performance. (NA?)

  • Consider utilizing a wide range of methods, datasets, and evaluation measures to ensure a comprehensive and unbiased assessment of the predictive performance of multi-label learning methods. (NA?)

  • Utilise a “tree-guided group lasso” methodology for multi-task regression problems involving structured sparsity, as it allows for a more accurate identification of shared covariates among related outputs. (NA?)

  • Carefully select the appropriate loss function when using gradient boosting machines (GBMs) for your specific data-driven task, as this choice significantly impacts the models performance and interpretability.’ (NA?)

  • Consider the dependence distribution, rather than solely focusing on individual dependencies, when evaluating the effectiveness of naive Bayes classifiers. (NA?)

  • Carefully evaluate and choose among multiple strategies for handling class imbalances in datasets, including data sampling, algorithmic modifications, and cost-sensitive learning, while also considering potential confounding factors like small disjuncts, lack of density and information, overlapping classes, noisy data, borderline instances, and dataset shifts. (NA?)

  • Carefully balance model complexity with the complexity of the underlying data to achieve optimal generalization, avoiding both underfitting and overfitting. (NA?)

  • Utilise the EUSBoost algorithm, which employs evolutionary undersampling guided boosting, to effectively handle highly imbalanced data sets in classification tasks. (NA?)

  • Utilise advanced machine learning techniques, such as random forests and approximate Gaussian processes, to improve the accuracy and scalability of runtime prediction models for complex algorithms. (NA?)

  • Consider using the proposed multi-task large margin nearest neighbor (mt-lmnn) algorithm for multi-task learning scenarios, as it effectively balances the importance of shared and task-specific parameters, leading to improved classification performance compared to existing methods. (NA?)

  • Carefully choose the appropriate F1 measure variant based on the relative importance they place on performance across different labels, as different choices can significantly affect the optimal predictions. (NA?)

  • Utilise Support Vector Regression (SVR) due to its ability to balance model complexity and prediction error through the use of an epsilon-insensitive loss function, providing a robust and accurate means of estimating continuous-valued functions. (NA?)

  • Focus on developing machine learning techniques specifically tailored for medical scoring systems, rather than relying on traditional methods that may compromise accuracy and sparsity. (NA?)

  • Use the proposed “initial adjustments” procedure to effectively initialize the solution of the minimization problem (6) before adding a new sample (x_new, y_new) into T, thereby improving the efficiency of the incremental v-SVR learning process. (NA?)

  • Focus on developing intelligible models that balance accuracy and interpretability, especially in mission-critical applications such as healthcare, where understanding the underlying mechanisms and potential biases is crucial for safe and effective implementation. (NA?)

  • Compare your proposed boosting algorithm (AdaBoost) against existing techniques like bagging, using a variety of weak learning algorithms and datasets, to demonstrate its superiority in reducing error rates and improving overall model performance. (NA?)

  • Evaluate the impact of feature selection on classifier security against evasion attacks before applying it to security-sensitive tasks. (NA?)

  • Consider adopting a nonparametric approach to generate very short-term predictive densities for renewable energy forecasting, particularly for solar power generation, as the distribution of forecast errors do not follow any of the common parametric densities. (NA?)

  • Adopt an “honest” approach to estimation, whereby one sample is used to construct the partition and another to estimate treatment effects for each subpopulation, enabling the construction of valid confidence intervals for treatment effects even with many covariates relative to the sample size, and without “sparsity” assumptions. (NA?)

  • Carefully choose appropriate study designs, ensure quality data collection and pre-processing, and utilize suitable machine learning algorithms to effectively analyze big datasets in order to accurately predict outcomes and gain valuable insights. (NA?)

  • Prioritize out-of-sample prediction as the primary metric for evaluating the efficacy of statistical learning algorithms, while remaining vigilant against potential pitfalls such as overfitting and ensuring that the chosen algorithm aligns with the specific goals of the study. (NA?)

  • Consider applying the Extreme Gradient Boosting (XGBoost) algorithm to analyze fMRI data in order to effectively classify patients with epilepsy from healthy individuals based on your language network patterns. (NA?)

  • Consider employing non-linear methods like gradient boosting machines for drug-target interaction prediction, as they can capture complex dependencies in the training data and generate prediction intervals for increased confidence in the results. (NA?)

  • Utilize probabilistic machine learning techniques, specifically Gaussian Process Regression, to infer solutions of differential equations using noisy multi-fidelity data, thereby enabling better understanding of uncertainty and facilitating adaptive solution refinement. (NA?)

  • Use a novel cost-sensitive boosting framework called “LinkBoost” for community-level network link prediction, which effectively handles the inherent skewness of network data and consistently performs as good as or better than many existing methods across multiple real-world network datasets. (NA?)

  • Employ a mixture model combining linear regression on bids with observable winning prices and censored regression on bids with censored winning prices, weighted by the winning rate of the DSP, to effectively handle the issue of censored data in real-time bidding systems. (NA?)

  • Consider using advanced undersampling techniques, such as evolutionary undersampling, undersampling by cleaning data, ensemble-based undersampling, and clustering-based undersampling, to effectively handle imbalanced datasets in various domains. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Consider using the proposed multi-task large margin nearest neighbor (mt-lmnn) algorithm for multi-task learning scenarios, as it effectively balances the importance of shared and task-specific parameters, leading to improved classification performance compared to existing methods. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Integrate multiple sources of information, such as miRNA functional similarity, disease semantic similarity, and known miRNA-disease associations, to create an informative feature vector for accurate prediction of miRNA-disease associations using advanced machine learning techniques like Extreme Gradient Boosting Machines. (NA?)

  • Consider using quantile regression instead of traditional mean regression when they are interested in estimating specific percentiles of a dependent variable, as it allows for a more comprehensive understanding of the underlying distribution. (NA?)

  • Focus on understanding the properties of the marginal likelihood function in order to optimize the performance of sparse Bayesian learning methods. (NA?)

  • Understand the relationship between various evaluation metrics and your underlying principles, such as precision and cost-weighted differences, in order to choose the most suitable metric for your specific application. (NA?)

  • Focus on using algorithmic experimentation to explore various machine learning methods through practical examples, while also considering potential limitations like the curse of dimensionality. (NA?)

  • Prioritise calibration alongside discrimination when developing and validating predictive algorithms, ensuring that the model accurately reflects the true probability of outcomes, thereby reducing potential harms associated with misleading predictions. (NA?)

  • Carefully choose the appropriate supervised machine learning algorithm for your disease prediction studies based on the relative performance of different algorithms, as demonstrated by the studycomparison of the Support Vector Machine (SVM), Naive Bayes, and Random Forest (RF) algorithms. (NA?)

  • Thoroughly analyze and optimize the hyperparameters of XGBoost, random forest, and gradient boosting models to ensure optimal performance across various datasets and tasks. (NA?)

  • Carefully choose appropriate evaluation metrics for your binary classification models, considering factors like class balance and interpretability, and avoid relying solely on commonly used measures like accuracy and F1 score without understanding your limitations. (NA?)

  • Employ a hybrid PCA-firefly algorithm for dimensionality reduction before applying the XGBoost algorithm for classification in intrusion detection systems. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Consider utilising a combination of XGBoost machine learning techniques and a clinically operable decision tree to develop a highly accurate and interpretable model for predicting COVID-19 patient mortality rates up to ten days in advance. (NA?)

  • Carefully choose appropriate evaluation metrics for binary classification problems, considering factors like prevalence, bias, and the relationship between the metrics themselves, to ensure accurate and meaningful interpretation of model performance. (NA?)

  • Carefully consider the possibility of omitted interaction bias when estimating treatment effect heterogeneity, and adopt appropriate techniques like post-double selection to minimize its impact. (NA?)

  • Utilize the novel (R^{*}) metric, which employs machine learning classifiers to assess Markov Chain Monte Carlo (MCMC) convergence, providing a comprehensive view of the entire joint distribution and offering improved detection of non-convergent chains compared to traditional methods like (). (NA?)

  • Use the alternating direction method of multipliers (ADMoM) to develop fully distributed training algorithms for support vector machines (SVMs) that are provably convergent to the centralized SVM, without requiring a central processing unit or exchanging training data among nodes. (NA?)

  • Utilise the Nystrom method for approximating a Gram matrix to improve kernel-based learning efficiency, particularly when dealing with large datasets. (NA?)

  • Utilise the restricted eigenvalue (RE) condition when working with high-dimensional linear regression problems, as it offers a less stringent requirement compared to other conditions like the restricted isometry property (RIP) and an earlier set of restricted eigenvalue conditions. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Be cautious when relying solely on the UCI repository for benchmarking purposes, as its datasets tend to be resistant to overfitting, leading to potentially misleading conclusions regarding the performance of various algorithms. (NA?)

Unsupervised Learning Algorithms

  • Focus on developing efficient algorithms for designing models and making accurate predictions while maintaining computational efficiency and robustness against noise in the context of big data. (Bosen Zhang et al. 2023)

  • Consider using a novel contrastive learning approach, ToThePoint, for efficient self-supervised learning of 3D point clouds, which involves recycling discarded features from the max-pooling operation and integrating them into the learning process, resulting in improved performance and reduced training time. (Xinglin Li et al. 2023)

  • Carefully choose appropriate pretext tasks, optimize hyperparameters, and utilize effective evaluation metrics to ensure successful implementation of self-supervised learning methods. (Balestriero et al. 2023)

  • Consider using a Prompt Ensemble Self-training (PEST) technique for open-vocabulary domain adaptation (OVDA) tasks, which leverages the synergy between vision and language to mitigate domain discrepancies in image and text distributions simultaneously, enabling effective learning of image-text correspondences in unlabeled target domains. (Jiaxing Huang et al. 2023)

  • Utilize semantic entropy - a novel entropy-based uncertainty measure that employs an algorithm for marginalizing over semantically-equivalent samples - to effectively estimate uncertainty in natural language processing tasks. (Kuhn, Gal, and Farquhar 2023)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Mollá 2023)

  • Adopt the POUF (Prompt-Oriented Unsupervised Fine-Tuning) technique when working with large pre-trained models. This involves directly fine-tuning the model or prompt on unlabelled target data, thereby improving the models ability to adapt to downstream tasks without requiring labeled data. (Tanwisuth et al. 2023)

  • Carefully consider the effects of self-supervision and contrastive alignment in deep multi-view clustering, as these factors can significantly impact cluster separability and overall performance, particularly when dealing with larger numbers of views. (Trosten et al. 2023)

  • Consider combining self-supervised contrastive learning with few-shot label information to improve graph anomaly detection performance, especially in cases where obtaining labeled anomaly data is challenging. (F. Xu et al. 2023)

  • Utilise variational Bayesian methods to evaluate the sensitivity of your conclusions to the choice of concentration parameter and stick-breaking distribution for inferences under Dirichlet process mixtures and related mixture models. (Giordano et al. 2023)

  • Utilise a novel Bayesian nonparametric method combining Markov random field models and mixture of finite mixtures models to analyse spatial income Lorenz curves, enabling simultaneous estimation of the number of clusters and the clustering configuration while taking into account geographical information. (G. Hu et al. 2023)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Kurian et al. 2023)

  • Utilize a combination of diverse top-k parameters for forming initial positive pairs during data augmentation, and implement a boundary distance constraint to accurately judge positive and negative relationships within mini-batches. This will significantly increase the robustness of your training processes. (Zhenhe Wu et al. 2023)

  • Carefully consider the effects of self-supervision and contrastive alignment in deep multi-view clustering, as these factors can significantly impact cluster separability and overall performance, particularly when dealing with larger numbers of views. (Hansen et al. 2023)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Sanderson 2023)

  • Consider utilizing a weak-supervision system called Osprey, which employs a generative modeling approach to estimate the accuracies and correlations of various labeling functions, ultimately combining these labels to produce probabilistic (confidence-weighted) training labels. (Kammoun et al. 2022)

  • Carefully consider the positive pairs they choose for contrastive learning, as selecting appropriate positive pairs can help avoid false positives and increase the variance of crops, leading to improved performance in various downstream tasks. (X. Peng et al. 2022)

  • Focus on developing methods that can effectively utilize unlabeled data of unknown class distributions, such as the adaptive consistency regularizer (ACR) proposed in the study, which dynamically estimates the true class distribution of unlabeled data and refines pseudo-labels accordingly. (Rizve, Kardan, and Shah 2022)

  • Utilize a novel unsupervised point cloud pre-training framework called “ProposalContrast” for 3D object detection, which learns robust 3D representations by contrasting region proposals, thereby improving the generalizability and transferability of your models. (J. Yin et al. 2022)

  • Utilize a novel sparse latent factor regression model to integrate heterogeneous large datasets, providing a tool for data exploration through dimensionality reduction and sparse low-rank covariance estimation while correcting for various batch effects. (Avalos-Pacheco, Rossell, and Savage 2022)

  • Consider using the matrix spike-and-slab LASSO prior for modeling joint sparsity in sparse spiked covariance models, as it offers rate-optimal posterior contraction for both the entire covariance matrix and the principal subspace, while also providing a point estimator with a rate-optimal risk bound. (F. Xie et al. 2022)

  • Utilize finite mixtures of exponential family random graph models (ERGMs) to effectively analyze and understand ensembles of networks, even in the presence of dyadic dependence and cross-graph heterogeneity. (F. Yin, Shen, and Butts 2022)

  • Employ an interactive contrastive learning model for self-supervised entity alignment, which involves creating pseudo-aligned entity pairs as pivots to facilitate direct cross-knowledge graph information interaction, integrating both textual and structural information, and carefully designing encoders for optimal utilisation in the self-supervised context. (K. Zeng et al. 2022)

  • Consider using a hash-like method for log parsing, which improves both robustness and efficiency compared to traditional tree-based methods. (Shijie Zhang and Wu 2021)

  • Consider implementing a novel latent contrastive learning (LaCoL) technique when dealing with noisy data in deep neural networks, as it enables the discovery of negative correlations within the data, thereby improving the overall robustness and generalization capabilities of the model. (Y. Bai et al. 2021)

  • Consider using the ARB (Align Representations with Base) approach in self-supervised learning, which involves maximizing the consistency between intermediate variables and representations of each view, leading to improved efficiency, reduced feature redundancy, and increased robustness to output dimension size compared to traditional symmetric contrastive learning methods. (Bardes, Ponce, and LeCun 2021)

  • Consider using Centered Kernel Alignment (CKA) to compare neural representations across different learning methods, such as self-supervised and supervised learning, to better understand the underlying mechanisms driving your performance differences. (Grigg et al. 2021)

  • Consider incorporating class relationship embedded similarity (CRS) into your contrastive learning processes, as it allows for more accurate expression of sample relationships in the output space and leads to improved performance in various domain adaptation tasks. (Junjie Li et al. 2021)

  • Employ Curriculum Pseudo Labeling (CPL) in semi-supervised learning (SSL) models to dynamically adjust thresholds based on the models learning status for each class, leading to improved accuracy and faster convergence.’ (Rizve et al. 2021)

  • Consider incorporating spatial consistency in your representation learning algorithms, especially for multi-object and location-specific tasks like object detection and instance segmentation, as it can improve the performance of fine-tuned models on various downstream localization tasks. (Roh et al. 2021)

  • Consider incorporating bounding boxes into pretraining processes to align convolutional features with foreground regions, thereby improving localization abilities and ultimately yielding superior transfer learning results for object detection. (Ceyuan Yang et al. 2021)

  • Utilize the semi-hierarchical Dirichlet process (semi-HDP) prior to avoid degeneracy issues associated with nested Dirichlet processes (NDP) and enable the identification of homogenous groups within heterogeneous populations. (Beraha, Guglielmi, and Quintana 2021)

  • Utilize a Bayesian tensor response regression (TRR) model with a multiway stick breaking shrinkage prior to analyze complex datasets with tensor-valued responses and scalar predictors, allowing for improved estimation accuracy and uncertainty quantification. (Guhaniyogi and Spencer 2021)

  • Utilize a hybrid mining method combining rough set theory and fuzzy set theory to improve efficiency and accuracy in generating association rules from large datasets. (R. Chatterjee et al. 2021)

  • Utilise a multi-task framework combining a supervised objective using ground-truth labels and a self-supervised objective reliant on clustering assignments with a single cross-entropy loss to achieve high-performance semi-supervised learning. (Assran et al. 2020)

  • Consider implementing a class-rebalancing self-training framework (CReST) to improve the performance of semi-supervised learning algorithms on class-imbalanced data. (Calderon-Ramirez et al. 2020)

  • Focus on achieving category-level alignment rather than instance-level alignment when dealing with partial view-alignment problems, as it offers higher accessibility and scalability for clustering and classification tasks. (Ting Chen et al. 2020)

  • Use entropy regularization to measure the dependency between learned features and class labels, thereby ensuring the conditional invariance of learned features and improving the generalization capabilities of your classifiers. (T. Fang et al. 2020)

  • Consider using a self-supervised image rotation task to evaluate the quality of your learned representations, as it shows a high rank correlation (>0.94) with traditional supervised evaluations, allowing them to effectively guide your unsupervised training processes without needing labeled data. (C. J. Reed et al. 2020)

  • Carefully examine the interplay between the number of negative samples, temperature, and margin parameters in your contrastive learning models, as these factors can significantly impact the performance of the model. (B. Zhu et al. 2020)

  • Carefully examine the interaction between data augmentation techniques and pre-training methods, as stronger data augmentation may negate the need for pre-training or even lead to worse performance, whereas self-training remains beneficial regardless of data augmentation strength. (Zoph et al. 2020)

  • Consider the differences between traditional statistical modeling and machine learning approaches, specifically regarding model interpretability and complexity, when choosing appropriate methods for your studies. (Badillo et al. 2020)

  • Develop an incremental version of the Centroid Decomposition technique to effectively recover multiple time series streams in linear time, thereby reducing the complexity from quadratic to linear and enabling accurate recovery of missing blocks in a continuous manner. (Khayati, Arous, et al. 2020)

  • Consider utilizing weak supervision approaches, such as Snorkel DryBell, to efficiently leverage diverse organizational knowledge resources for training high-quality machine learning models without requiring extensive manual data labeling efforts. (“Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence” 2019)

  • Consider using the “Bag of Instances Aggregation” (BINGO) approach when working with self-supervised learning, particularly for small-scale models, as it enables efficient transfer of relationships among similar samples, leading to improved performance. (“Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence” 2019)

  • Utilise knowledge distillation (KD) rather than adversarial domain adaptation (ADA) for semi-supervised domain adaptation of deep neural networks (DNNs) because KD doesnt necessitate dataset-specific hyperparameter tuning, thus being universally applicable. (Orbes-Arteaga et al. 2019)

  • Utilize a combination of feature whitening and consensus loss in unsupervised domain adaptation to improve the accuracy of your models across multiple datasets. (S. Roy et al. 2019)

  • Carefully consider the domain of unlabelled data used for self-supervision in few-shot learning scenarios, as selecting images from a similar domain can greatly enhance performance, whereas using images from a different domain could negatively impact it. (J.-C. Su, Maji, and Hariharan 2019)

  • Utilize a robust PCA-based algorithm for learning dependency structures in weak supervision models, which can lead to improved theoretical recovery rates and outperform existing methods on various real-world tasks. (Varma et al. 2019)

  • Consider using local aggregation (LA) for unsupervised learning of visual embeddings, which involves training an embedding function to maximize a metric of local aggregation, causing similar data instances to move together in the embedding space while allowing dissimilar instances to separate, thereby enabling effective unsupervised transfer learning performance on various large-scale visual recognition datasets. (C. Zhuang, Zhai, and Yamins 2019)

  • Consider using Monte Carlo simulation methods to generate controlled datasets for evaluating the performance of algorithms in handling class imbalance issues in machine learning tasks. (Abdar et al. 2019)

  • Utilise the Rlda package for mixed-membership clustering analysis, especially when dealing with various types of categorical data like Multinomial, Bernoulli, and Binomial entries. This package offers a unique Bayesian LDA model that allows for the selection of the optimal number of clusters based on a truncated stick-breaking prior approach, thereby providing regularisation of model results. (Albuquerque, Valle, and Li 2019)

  • Consider using the M-GRAF model when analyzing multiple binary networks with similar patterns, as it allows for the extraction of both common and low-dimensional individual-specific structure, leading to improved prediction and understanding of individual variations in human cognitive traits and behaviors. (Lu Wang, Zhang, and Dunson 2019)

  • Utilise the proposed ISG+D-Spot methodology for accurate and efficient detection of fraudulent entities in multidimensional data, particularly when dealing with hidden-densest blocks. (Yikun et al. 2019)

  • Consider utilizing unsupervised prompt tuning techniques such as Nested Mean Teaching and Dual Complementary Teaching when working with text-driven object detection systems, as these approaches can significantly enhance performance without requiring manual annotations. (Devlin et al. 2018)

  • Carefully consider the tradeoffs between precision, dimensionality, and graph properties when working with hyperbolic embeddings, as well as explore alternative optimization strategies such as adding a learnable scale term or utilizing Stochastic Gradient Descent-based algorithms to improve the quality of embeddings. (Sa et al. 2018)

  • Utilize a Ward-like hierarchical clustering algorithm that includes spatial/geographical constraints through the use of two dissimilarity matrices, allowing them to balance the tradeoff between increasing spatial contiguity and maintaining the quality of the solution based on the variables of interest. (Chavent et al. 2018)

  • Utilise a computer-assisted algorithm to discover keywords and document sets from unstructured text, thereby improving the efficiency and effectiveness of your analyses. (G. King, Lam, and Roberts 2017)

  • Consider utilizing weak supervision methods, such as those provided by Snorkel, to efficiently generate large amounts of training data for machine learning models without requiring extensive manual labeling efforts. (Dehghani et al. 2017)

  • Carefully consider the appropriate fusion of local and global graph structure information when conducting multi-view clustering on graph data. (G. Ma et al. 2017)

  • Use a combination of multiple cluster validity indices to improve the accuracy of identifying natural clusters in acoustic emission signals, rather than relying on just one index. (Jialin Tang et al. 2017)

  • Avoid making assumptions of independence between variables during the variable selection process for latent class analysis, as doing so can lead to incorrect conclusions about the relevance of variables for clustering. (Fop, Smart, and Murphy 2017)

  • Utilize the Wasserstein metric to provide pseudo labels for unlabeled images in a semi-supervised learning context for image classification tasks. (Arjovsky, Chintala, and Bottou 2017)

  • Consider using the MeanShift++ algorithm for mode-seeking clustering tasks, especially in low-dimensional applications like image segmentation and object tracking, as it offers significant improvements in speed without compromising clustering quality. (Bigdeli and Zwicker 2017)

  • Consider using co-regularized domain alignment for unsupervised domain adaptation, which involves constructing multiple diverse feature spaces and aligning source and target distributions within each space, while ensuring that the alignments agree with each other regarding class predictions on unlabeled target examples. (Bousmalis et al. 2017)

  • Consider using non-parametric instance discrimination for unsupervised feature learning, as it enables the learning of a good feature representation that captures apparent similarity among instances, leading to improved performance in various tasks such as image classification, semi-supervised learning, and object detection. (Doersch and Zisserman 2017)

  • Consider using a combination of instance-level and graph-level matching for assignment and feature learning, respectively, in order to achieve more stable and superior results in semi-supervised learning. (Priya Goyal et al. 2017)

  • Carefully consider the use of semi-supervised learning methods when dealing with limited labeled data, as these techniques can effectively leverage unlabeled data to improve classification performance while minimizing potential risks such as asymptotic bias. (Laine and Aila 2016)

  • Consider jointly optimizing dimensionality reduction and clustering tasks, particularly when working with nonlinear transformations, to achieve improved clustering outcomes. (Bo Yang et al. 2016)

  • Consider using the Wasserstein dependency measure instead of mutual information maximization for representation learning, especially in situations where the mutual information is large, as it provides more robust and comprehensive representations. (Alain and Bengio 2016)

  • Focus on understanding and exploiting the unique characteristics of deep learning workloads, such as feedback-driven exploration, heterogeneity, and intra-job predictability, to develop specialized scheduling frameworks that can improve latency and efficiency in training deep learning models. (Tianqi Chen et al. 2016)

  • Focus on developing unsupervised learning algorithms that mimic the way humans naturally process visual information, specifically by leveraging motion-based grouping cues to learn effective visual representations. (Pathak et al. 2016)

  • Aim to maximise the information between data indices and labels while explicitly enforcing the equipartition condition, which helps avoid degenerate solutions and improve the quality of unsupervised learning. (Dosovitskiy et al. 2016)

  • Utilize a generative model for mining sequential patterns in databases, specifically one that involves iteratively sampling subsequences from a set of interesting sequences and randomly interleaving them to form the database sequence. (Fowkes and Sutton 2016)

  • Utilise an end-to-end framework, specifically Log-Mine’, which offers an unsupervised, quick, and memory-efficient solution for processing vast amounts of log messages through a hierarchical pattern recognition system. (Hamooni et al. 2016)

  • Consider utilizing self-ensembling for visual domain adaptation problems, specifically by modifying the mean teacher variant of temporal ensembling, as it has been proven to achieve state-of-the-art results in various benchmarks and even surpass the performance of traditional supervised learning in certain cases. (Yanghao Li et al. 2016)

  • Consider using A-tSNE, a novel approach to adapt the complete tSNE pipeline for progressive visual analytics, which significantly reduces initialization time and allows for interactive modification, removal, or addition of high-dimensional data without disrupting the visual analysis process. (Pezzotti et al. 2015)

  • Focus on developing simple, efficient, and effective unsupervised domain adaptation methods like CORAL, which aligns the second-order statistics of source and target distributions without requiring any target labels, leading to improved performance in various application areas. (B. Sun, Feng, and Saenko 2015)

  • Utilise a novel model-based clustering method specifically tailored for time series data, called FunFEM, to analyse and compare multiple European Bike Sharing Systems (BSSs). (Bouveyron, Côme, and Jacques 2015)

  • Consider applying the redundancy-reduction principle to self-supervised learning, as demonstrated by the success of the Barlow Twins method in achieving state-of-the-art results on various computer vision tasks. (T. T. Cai, Liang, and Zhou 2015)

  • Carefully choose the right distance measure for your specific time-series clustering task, as it can greatly impact the accuracy and efficiency of the clustering process. (Paparrizos and Gravano 2015)

  • Consider maximizing representation entanglement by incorporating a bonus proportional to the soft nearest neighbor loss into your training objective, as it acts as a regularizer and improves handling of outlier data. (Azadi et al. 2015)

  • Consider developing self-supervised learning methods for 3D data that remain agnostic to the underlying neural network architecture and specifically leverage the geometric nature of 3D point cloud data, leading to improved transfer learning and better performance on downstream applications. (A. X. Chang et al. 2015)

  • Consider using a Bagged Outlier Representation Ensemble (BORE) for outlier detection, which combines unsupervised outlier scoring functions (OSFs) as features in a supervised learning framework, allowing for adaptation to arbitrary OSF feature representations, class imbalance, and prediction-time constraints on computational cost. (Micenková, McWilliams, and Assent 2015)

  • Utilise a combination of nuclear-norm-regularised matrix approximation and maximum-margin matrix factorisation techniques when tackling matrix-completion problems, resulting in improved efficiency and accuracy. (Hastie et al. 2014)

  • Utilise the FFDiag algorithm for fast and efficient joint diagonalisation of multiple matrices, particularly in situations where orthogonality cannot be assumed. (Tichavsky, Phan, and Cichocki 2014)

  • Consider integrating content information into the group modeling process to improve the efficiency and accuracy of spammer detection algorithms. (Low et al. 2014)

  • Utilize the Odd Sketch methodology for estimating the Jaccard similarity of two sets, as it effectively reduces the variance when the similarity is close to 1 compared to traditional methods like minwise hashing. (Mitzenmacher, Pagh, and Pham 2014)

  • Utilise a novel dissimilarity-based sparse subset selection (DS3) algorithm for identifying optimal representatives within large collections of data points or models. This algorithm offers numerous benefits over previous approaches including scalability, flexibility in handling various types of dissimilarities, robustness against outliers, and ability to handle multiple groups within the data. (Elhamifar, Sapiro, and Sastry 2014)

  • Focus on developing a reliable density estimation algorithm based on local connectivity between K nearest neighbors (KNN) to effectively exclude negative pairs from the KNN graph while maintaining sufficient positive pairs, leading to improved clustering performance. (D. Yi et al. 2014)

  • Consider utilizing unlabelled data when working with limited labelled samples, as demonstrated through the success of various approaches in the two machine learning contests discussed. (I. J. Goodfellow, Erhan, et al. 2013)

  • Avoid making unnecessary assumptions about the underlying distribution of continuous variables in Bayesian networks, and instead utilize nonparametric density estimation techniques like kernel density estimation to achieve greater accuracy in modeling complex relationships. (John and Langley 2013)

  • Carefully consider the choice of initialization scheme when applying the EM algorithm for clustering in high dimensions, as it can greatly impact the final solution quality. (Meila and Heckerman 2013)

  • Integrate a computational algorithm called Topic Rose Tree with an interactive visual interface to create a visual analytics system called HierarchicalTopics (HT), which helps users navigate and understand large text collections by organizing topics into a hierarchical structure and providing temporal evolution views. (W. Dou et al. 2013)

  • Utilise adversarial domain adaptation techniques to discover and control for latent confounds in text classification, thus enhancing the robustness of your models against confounding shift. (Diederik P. Kingma and Welling 2013)

  • Utilise tensor decompositions for learning latent variable models, as it allows for computationally and statistically efficient parameter estimation through the extraction of a certain orthogonal decomposition of a symmetric tensor derived from the observable moments. (Anima Anandkumar et al. 2012)

  • Utilise a novel method of moments approach for parameter estimation in high-dimensional mixture models and hidden Markov models, which is computationally efficient, based on low-order moments, and provides unsupervised learning guarantees under mild rank conditions. (Animashree Anandkumar, Hsu, and Kakade 2012)

  • Utilize Bayesian rose trees instead of traditional binary trees for hierarchical clustering tasks, as they provide a richer representation of the underlying data structure and lead to more accurate and interpretable results. (Blundell, Teh, and Heller 2012)

  • Avoid relying solely on multi-objective optimization with predefined norms for recovering simultaneously structured models, as it offers no improvement over algorithms that exploit just one structure, and instead explore novel convex relaxations tailored specifically to the multiple structures involved. (Oymak et al. 2012)

  • Optimize your models for the appropriate criterion, rather than simply applying existing techniques without considering whether they are best suited to the task at hand. (Rendle et al. 2012)

  • Consider utilizing advanced techniques such as spatiotemporal modeling, functional data analysis, and kriging when analyzing complex datasets involving both spatial and temporal dependencies, rather than simply applying traditional statistical methods. (Gromenko et al. 2012)

  • Explore the potential of integrating Bayesian nonparametric methods with traditional hard clustering algorithms, such as k-means, to develop more efficient and effective clustering solutions. (Kulis and Jordan 2011)

  • Utilize a novel visualization tool to navigate the vast landscape of potential clusterings, allowing them to efficiently identify and select the most appropriate clustering solution for your specific research goals. (Grimmer and King 2011)

  • Focus on creating a unified, feature-based matrix factorization model that can accommodate diverse types of information, rather than designing separate models for each type of information. (Tianqi Chen et al. 2011)

  • Focus on developing unsupervised techniques for extracting product attributes and your values from e-commerce product pages, rather than relying on distant supervision or manual annotation, due to the limitations of existing knowledge bases and the diversity of product types. (“Advances in Information Retrieval” 2009)

  • Utilise the LAS algorithm, a statistically motivated biclustering procedure, to identify large average submatrices within a given real-valued data matrix. This process operates iteratively, balancing the trade-off between the size of a submatrix and its average value, and is connected to the minimum description length principle. (Shabalin et al. 2009)

  • Utilize the OptSpace algorithm for matrix completion tasks, particularly when dealing with approximately low-rank matrices, due to its order-optimal performance guarantees in various scenarios. (J.-F. Cai, Candes, and Shen 2008)

  • Use Bayesian nonnegative matrix factorization (NMF) for community detection tasks, as it provides overlapping or soft-partitioning solutions, soft-membership distributions, excellent module identification capabilities, and avoids the drawbacks of modularity optimization methods like the resolution limit. (Heinson 2008)

  • Consider using separate ranking losses for labeled and unlabeled data sets in your analysis, rather than combining them, to improve the accuracy of your models. (M. R. Amini, Truong, and Goutte 2008)

  • Understand the differences between the unnormalized graph Laplacian, the normalized graph Laplacian according to Shi and Malik (2000), and the normalized graph Laplacian according to Ng, Jordan, and Weiss (2002) when implementing spectral clustering algorithms, as these variations impact the performance and interpretation of the clustering results. (Luxburg 2007)

  • Understand the underlying principles of spectral clustering algorithms, including the differences between unnormalized and normalized graph Laplacians, and choose the appropriate algorithm based on your specific application and dataset characteristics. (Luxburg 2007)

  • Utilize Bayesian methods for density regression, specifically employing a nonparametric mixture of regression models, to effectively capture the complex relationship between a random probability distribution and multiple predictors. (Dunson, Pillai, and Park 2007)

  • Consider implementing distributed algorithms for topic models, specifically Latent Dirichlet Allocation (LDA) and Hierarchical Dirichlet Process (HDP) models, to efficiently handle large datasets while maintaining high accuracy in your analyses. (A. S. Das et al. 2007)

  • Consider employing a nonparametric Bayesian approach when analyzing microarray data to detect differentially expressed genes, as it offers several advantages over existing methods, such as providing a full description of uncertainties, enabling inference without a null sample, and allowing for joint inference on multiple genes. (Lewin, Bochkina, and Richardson 2007)

  • Carefully evaluate the appropriateness of predictive accuracy as a performance measure when dealing with imbalanced datasets, and consider alternative metrics like ROC curves, precision and recall, and cost-sensitive measures. (“Data Mining and Knowledge Discovery Handbook” 2005)

  • Adopt a hierarchical statistical modelling framework for performing areal wombling, allowing for direct estimation of the probability that two geographic regions are separated by the wombled boundary, and enabling accurate estimation of quantities that would otherwise be inestimable using classical approaches. (Haolan Lu and Carlin 2005)

  • Consider utilising the aids (Automatic Distillation of Structure) algorithm for grammar-like rule induction, which combines statistics and rules, and is able to discover hierarchical structure in any sequence data based on the minimal assumption that the corpus at hand contains partially overlapping strings at multiple levels of organisation. (Solan et al. 2005)

  • Employ latent factor regression models to address the challenges posed by the large p, small n’ paradigm, specifically in areas like gene expression analysis. (“Bayesian Statistics 7” 2003)

  • Utilize a Bayesian nonparametric approach for analyzing spatial count data, specifically extending the Bayesian partition methodology to handle count data, allowing for probability statements on incidence rates around point sources without making any parametric assumptions about the nature of the influence between the sources and the surrounding location. (Denison and Holmes 2001)

  • Utilize a Bayesian approach to classification problems, which allows for the incorporation of prior knowledge and the balancing of model complexity against fit to the data, leading to improved performance compared to traditional maximum likelihood methods. (Hand and Yu 2001)

  • Consider adopting a top-down induction of clustering trees approach, which combines principles from instance-based learning and decision tree induction, to effectively identify clusters in various types of data. (Blockeel, Raedt, and Ramon 2000)

  • Focus on identifying emerging patterns (EPs) with low to medium support (1%-20%) in order to gain valuable insights and guidance in various fields, as these EPs often provide new knowledge that cannot be easily discovered through traditional statistical methods. (G. Dong and Li 1999)

  • Use self-supervised learning techniques like self-prediction and contrastive learning to effectively extract meaningful patterns from large amounts of unlabelled data, enabling efficient knowledge transfer to various downstream tasks. (Yarowsky 1995)

  • Utilize Contrastive Predictive Coding (CPC) as an unsupervised objective for learning predictable representations, which can significantly enhance the data-efficiency of image recognition tasks. (Barlow 1989)

  • Carefully evaluate and choose suitable stopping rules for determining the number of clusters in a dataset, considering your performance and potential data dependency. (Milligan and Cooper 1985)

  • Consider implementing an asymmetric Dirichlet prior over the document-topic distributions in your LDA models, as it offers significant improvements in model performance and robustness without incurring additional computational costs. (Geman and Geman 1984)

  • Use the gSpan algorithm for efficient graph-based pattern mining, which employs depth-first search and DFS Lexicographic order to systematically explore and prune the search space without generating candidates, thereby reducing computational costs and increasing speed compared to traditional methods. (X. Yan and Han, n.d.)

  • Utilize weighted low-rank approximations for analyzing datasets with non-uniform sampling or noise levels, as it leads to more accurate representations of the underlying structures compared to traditional unweighted approaches. (NA?)

  • Carefully consider the choice of distance measure, clustering algorithm, and number of clusters when conducting clustering analysis, as these decisions significantly impact the resulting clusters and subsequent interpretations. (NA?)

  • Carefully examine and compare the properties of various objective measures before choosing the appropriate one for your specific application, taking into account factors like invariance under row and column scaling operations, sensitivity to support-based pruning, and consistency with domain expert expectations. (NA?)

  • Utilise the RCA algorithm for learning distance metrics using side-information in the form of groups of “similar” points, as it demonstrates superior efficiency and cost-effectiveness compared to alternatives while achieving comparable improvements in clustering performance. (NA?)

  • Carefully consider the type of data being analyzed, the efficiency and scalability of data mining algorithms, the usefulness and certainty of results, the expression of data mining requests and results, interactive mining at multiple abstraction levels, mining information from different sources, and protection of privacy and data security when developing data mining techniques. (NA?)

  • Develop flexible learning algorithms capable of adapting to concept drift and hidden contexts through techniques such as maintaining a window of trusted examples and hypotheses, storing and reusing concept descriptions, and monitoring system behavior via heuristics. (NA?)

  • Focus on developing incremental conceptual clustering algorithms that prioritize maximizing inference capabilities while being computationally efficient and flexible enough to apply across various domains. (NA?)

  • Recognize the unique challenges and opportunities associated with data mining, including dealing with massive datasets, handling contaminated data, addressing nonstationarity and selection biases, and effectively utilizing automated data analysis techniques while maintaining a focus on substantive significance. (NA?)

  • Carefully consider and develop appropriate data preparation techniques to accurately identify unique users, user sessions, and semantically meaningful transactions in order to effectively analyze and draw insights from web usage data. (NA?)

  • Consider utilizing unsupervised learning techniques, specifically one-class SVM, for seizure detection tasks, as it provides numerous benefits including eliminating the need for patient-specific tuning, reducing reliance on costly seizure data collection, and enabling accurate detection without requiring precise marking of seizure intervals. (NA?)

  • Utilize the Hilbert-Schmidt independence criterion (HSIC) test to assess the statistical significance of dependencies detected by kernel independence measures, particularly for multivariate data and structured data like texts. (NA?)

  • Use a novel definition of principal curves as continuous curves of a given length that minimize the expected squared distance between the curve and points of the space randomly chosen according to a given distribution, leading to improved theoretical analysis and practical construction. (NA?)

  • Utilise a novel approach for clustering categorical data based on an iterative method for assigning and propagating weights on the categorical values in a table, leading to a similarity measure arising from the co-occurrence of values in the dataset. (NA?)

  • Utilise a novel approach for clustering categorical data based on an iterative method for assigning and propagating weights on the categorical values in a table, leading to a similarity measure arising from the co-occurrence of values in the dataset. (NA?)

  • Carefully consider multiple properties of your chosen interestingness measure, including symmetry under variable permutation, row/column scaling invariance, and antisymmetry under row/column permutation, to ensure accurate and meaningful interpretation of association patterns in your dataset. (NA?)

  • Consider using a recursive unsupervised learning approach for estimating the parameters of finite mixture models, which allows for simultaneous selection of the optimal number of components in the model. (NA?)

  • Consider using machine learning techniques, specifically the EM clustering algorithm, to analyze and categorize packet header traces in network analysis, allowing them to identify patterns and trends in traffic behavior. (NA?)

  • Leverage the inherent geometry of your data to create representations, invariant maps, and learning algorithms that capture the low-dimensional structure of the data, allowing for improved classification performance. (NA?)

  • Use a novel optimization technique based on semidefinite programming to bridge the gap between kernel methods and manifold learning, allowing for more accurate detection of the dimensionality of underlying manifolds and discovery of your modes of variability. (NA?)

  • Utilise the novel algorithm presented, which efficiently solves nuclear norm regularised problems without requiring singular value decompositions, thus reducing computational complexity and increasing scalability. (NA?)

  • Consider utilizing generative model-based clustering approaches, particularly those based on von Mises-Fisher (vMF) distributions, due to your superior performance in certain scenarios and lower computational costs compared to some alternative methods. (NA?)

  • Utilize the Extended Motif Discovery (EMD) algorithm when dealing with multi-dimensional time-series data, as it allows for the extraction of both Same Length (SL) and Different Lengths (DL) patterns, thereby providing a more accurate and comprehensive understanding of the underlying data structure. (NA?)

  • Optimize a likelihood-type measure when developing algorithms for learning the structure of Markov logic networks (MLNs), rather than relying solely on off-the-shelf inductive logic programming (ILP) systems, as this leads to better performance and improved probabilistic predictions. (NA?)

  • Utilize a direct gradient-based optimization method for Maximum Margin Matrix Factorization (MMMF) in large collaborative prediction problems, as it demonstrates superior performance compared to existing methods. (NA?)

  • Utilise diffusion semigroups to create multi-scale geometries within complex structures, allowing for the organisation and representation of said structures through the selection of appropriate eigenfunctions or scaling functions of Markov matrices. (NA?)

  • Carefully consider the impact of various parameters, such as text segment length and stop-word inclusion, on the stability and reproducibility of the Leximancer-generated concept maps, ensuring that the chosen settings accurately capture the intended semantic relationships within the text. (NA?)

  • Consider utilizing Bregman divergences in your clustering algorithms, as it allows for improved performance and offers a connection to boosting techniques. (NA?)

  • Focus on selecting a good encoder rather than spending resources on training, as the choice of encoder plays a significant role in achieving superior performance in sparse coding and vector quantization applications. (NA?)

  • Carefully consider the potential impact of diagonal dominance on the performance of kernel-based clustering algorithms, especially when dealing with sparse high-dimensional data like text corpora, and explore various strategies to mitigate this issue, such as using subpolynomial kernels, diagonal shifts, or algorithmic modifications. (NA?)

  • Utilize the pachinko allocation model (PAM) instead of Latent Dirichlet allocation (LDA) for better representation and understanding of topic correlations in text analysis. (NA?)

  • Carefully consider the impact of design choices and parameter values when evaluating and comparing psychological models using word co-occurrence statistics for semantic representation. (NA?)

  • Utilise a Monte Carlo cross-entropy algorithm for weighted rank aggregation of cluster validation measures to effectively compare and evaluate the performance of different clustering algorithms. (NA?)

  • Consider using locally adaptive metrics for clustering high-dimensional data, rather than relying solely on global dimensionality reduction techniques, in order to effectively capture local correlations and improve overall performance. (NA?)

  • Consider utilizing a Bayesian approach combined with adaptive views clustering for improved 3-D model retrieval, particularly when dealing with large datasets. (NA?)

  • Consider both local and nonlocal quantities when developing unsupervised discriminant projection (UDP) techniques for dimensionality reduction of high-dimensional data in small sample size cases, as this approach allows for simultaneous maximization of nonlocal scatter and minimization of local scatter, resulting in improved performance compared to traditional methods. (NA?)

  • Utilise a co-clustering based classification (CoCC) algorithm to effectively transfer knowledge from in-domain data to out-of-domain data, thereby significantly improving classification performance in situations where labeled data is limited or absent in the target domain. (NA?)

  • Consider utilizing self-taught learning algorithms, which leverage unlabeled data to improve performance on supervised classification tasks, across various input modalities like images, audio, and text. (NA?)

  • Utilize the Singular Value Projection (SVP) algorithm for solving Affine Rank Minimization Problems (ARMP) due to its ability to guarantee geometric convergence rates, even in the presence of noise, and requiring less restrictive assumptions on Restricted Isometry Property (RIP) constants compared to other existing methods. Additionally, incorporating a Newton-step into the SVP framework can further enhance the efficiency and effectiveness of the algorithm. (NA?)

  • Use spectral clustering algorithms, specifically the Normalized Spectral Clustering Algorithm based on either the Symmetric Normalized Graph Laplacian or Random Walk Normalized Graph Laplacian, to effectively analyze complex datasets and improve clustering performance compared to traditional methods. (NA?)

  • Consider using Spectral Regression Discriminant Analysis (SRDA) instead of traditional Linear Discriminant Analysis (LDA) for large-scale datasets due to its superior computational efficiency and ability to handle regularization techniques. (NA?)

  • Consider implementing an iterative sampling procedure to enhance the precision of your results, particularly when dealing with complex datasets or models. (NA?)

  • Carefully consider how they manage discretization bias and variance in naive-Bayes learning, as proper management can significantly reduce classification errors. (NA?)

  • Consider utilizing equivalence constraints, particularly positive ones, in unsupervised learning tasks to improve the quality of your models and achieve better results. (NA?)

  • Extend the Hierarchical Dirichlet Process Hidden Markov Model (HDP-HMM) to include a parameter for self-transition bias and place a separate prior on this parameter to improve the models ability to handle state persistence and achieve better performance in tasks such as speaker diarization.’ (NA?)

  • Utilise the Support Vector Clustering (SVC) algorithm for effective clustering of data sets. This involves mapping data points onto a high dimensional feature space via a Gaussian kernel, searching for the minimum encompassing sphere within this space, and interpreting the resulting contours as cluster boundaries upon returning to the data space. The width of the Gaussian kernel and the soft margin constant control the scale at which the data is examined and help manage outliers and overlapping clusters, respectively (NA?)

  • Consider departing from the traditional Gaussianity assumption when working with continuous-valued data, as doing so enables the estimation of the full causal model rather than just a set of possible models. (NA?)

  • Utilize the k-modes algorithm for clustering large datasets with categorical values, as it effectively extends the k-means algorithm to categorical domains while maintaining efficiency. (NA?)

  • Utilize the concept of closed frequent itemsets’ when conducting association rule mining tasks because it significantly reduces the number of redundant rules produced while maintaining the exact frequency of all frequent itemsets. (NA?)

  • Utilize the quantics tensor method for approximating high-dimensional numerical models, as it offers near-optimal computational efficiency and avoids the curse of dimensionality’. (NA?)

  • Utilise a supervised learning approach with a modified loss function to achieve greater accuracy in discriminating between target and decoy peptide spectral matches (PSMs) in mass spectrometry analysis. (NA?)

  • Use the co-ranking matrix as a unifying framework to evaluate and compare the effectiveness of different dimensionality reduction algorithms, taking into consideration factors such as precision, recall, and overall quality. (NA?)

  • Use Labeled LDA, a supervised topic model that constrains Latent Dirichlet Allocation by defining a one-to-one correspondence between LDAs latent topics and user tags, allowing for direct learning of word-tag correspondences and improving credit attribution in multi-labeled corpora.’ (NA?)

  • Utilize the Dirichlet Forest model for topic modeling, which effectively incorporates domain knowledge via Must-Link and Cannot-Link primitives, resulting in improved accuracy and interpretability compared to traditional Latent Dirichlet Allocation models. (NA?)

  • Utilise multiple views of the data to relax stringent requirements needed for clustering algorithms to succeed, particularly when using Canonical Correlation Analysis (CCA) to project the data into a lower-dimensional subspace. (NA?)

  • Use alternative methods like Chib-style estimator and left-to-right evaluation algorithm instead of common methods like harmonic mean method and empirical likelihood method for accurately estimating the probability of held-out documents in topic modelling. (NA?)

  • Carefully choose the appropriate cluster concept (such as modality-based or pattern-based) depending on the specific application and requirements, and then utilize suitable methods for merging Gaussian mixture components accordingly. (NA?)

  • Utilise Non-negative Matrix Factorisation (NMF) based algorithms for community discovery in complex networks due to your high interpretability, ability to handle overlapping clusters, and ease of incorporating prior knowledge. (NA?)

  • Consider utilizing the clusterMaker plugin for Cytoscape, which offers a range of clustering algorithms and visualizations that can be employed individually or collectively for the examination and representation of biological datasets, as well as for validating or creating hypotheses regarding biological function. (NA?)

  • Employ generative probabilistic models for multi-label document classification, especially in large-scale corpora, because these models allow for explicit assignment of individual words to specific labels and simultaneous modeling of all labels, leading to improved handling of dependencies between labels. (NA?)

  • Consider utilizing equivalence constraints, particularly positive ones, in unsupervised learning tasks, as they can significantly improve the quality of the learned representation and enable better clustering and classification outcomes. (NA?)

  • Utilize a Bayesian method called Multiple Dataset Integration (MDI) for unsupervised integrative modeling of multiple datasets in order to efficiently combine information from various data types and improve the accuracy of your analysis. (NA?)

  • Utilise the “Score Matching” technique for estimating non-normalised statistical models, which involves minimising the expected squared distance between the gradient of the log-density given by the model and the gradient of the log-density of the observed data. (NA?)

  • Consider utilizing a semi-supervised hashing (SSH) framework for large-scale search tasks, which combines supervised empirical fitness and unsupervised information theoretic regularization to optimize the accuracy of hash functions while mitigating the risk of overfitting. (NA?)

  • Consider using a variant of the k-means clustering algorithm to minimize N-subjettiness, which improves the tagging performance of N-subjettiness for identifying boosted hadronic objects such as top quarks. (NA?)

  • Use a nonlinear successive over-relaxation (SOR) algorithm instead of a standard alternating minimization scheme for solving low-rank factorization models, as it provides significant improvements in speed and accuracy. (NA?)

  • Use Probabilistic Latent Semantic Analysis (PLSA) instead of Latent Semantic Analysis (LSA) because it provides a statistically sound foundation, well-defined probabilities, explicable results, and superior performance in tasks such as automatic indexing and handling polysemous words. (NA?)

  • Consider the various factors influencing self-labeled techniques for semi-supervised learning, such as addition mechanisms, single-classifier vs multi-classifier, single-learning vs multi-learning, and single-view vs multi-view, when selecting appropriate methods for your specific datasets and goals. (NA?)

  • Consider employing machine learning techniques, specifically latent variable modelling, to better understand the complex relationships between symptom transitions and identify patterns of symptoms within children, challenging the traditional atopic march’ paradigm.’ (NA?)

  • Utilize the Decoding Toolbox (TDT) for efficient, reliable, and flexible multivariate analysis of functional brain imaging data, enabling better sensitivity, specificity, and prediction of cognitive and mental states. (NA?)

  • Utilize VizBin, a Java-based application, for efficient and intuitive reference-independent visualization of metagenomic datasets from single samples, enabling human-in-the-loop inspection and binning, thereby improving the accuracy and reliability of metagenomic data analysis. (NA?)

  • Carefully select and evaluate the appropriate machine learning algorithm for your specific geomorphological problem, taking into account the type of data, desired outcome, and computational requirements. (NA?)

  • Leverage the low-rank property of certain matrices to develop efficient algorithms for recovering the full matrix from incomplete observations, thereby addressing the challenge posed by the impossibility of fully sampling large matrices. (NA?)

  • Consider utilizing tensor decomposition techniques for signal processing and machine learning tasks, as they offer advantages such as uniqueness and robustness compared to traditional matrix-based methods. (NA?)

  • Utilise a windowed technique to learn parsimonious time-varying autoregressive models from multivariate timeseries, modelling the stack of potentially different system matrices as a low rank tensor for improved interpretability and scalability. (NA?)

  • Consider using the YADING algorithm for fast and accurate clustering of large-scale time series data, which consists of three steps: sampling the input dataset, conducting clustering on the sampled dataset, and assigning the rest of the input data to the clusters generated on the sampled dataset. (NA?)

  • Consider developing a Hierarchical Importance-aware Factorization Machine (HIFM) for predicting response in mobile advertising, as it effectively addresses the challenges of temporal dynamics, cold-start issues, and the need for good regression and ranking performance. (NA?)

  • Carefully select and interpret the type of data fed into machine learning algorithms, as different forms of data can lead to complementary insights about the underlying physics. (NA?)

  • Consider using a multi-view low-rank sparse subspace clustering algorithm to learn a joint subspace representation by constructing an affinity matrix shared among all views, while balancing the agreement across different views and encouraging sparsity and low-rankness of the solution. (NA?)

  • Focus on developing a deep understanding of the underlying connections between different network embedding models, such as DeepWalk, LINE, PTE, and node2vec, in order to improve the efficiency and effectiveness of these models for various applications. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Utilise unsupervised machine learning techniques like diffusion maps to effectively classify topological phase transitions in complex systems without requiring any prior labelling or knowledge about the underlying phases. (NA?)

  • Carefully consider the type of Positive Unlabeled (PU) learning scenario they are dealing with - Single-Training-Set Scenario or Case-Control Scenario - as this affects the interpretation of results and choice of appropriate methods. (NA?)

  • Consider developing and utilising new distance metrics like advanced metric $d_{

exttt{AMA}}$ and extended metric $d_{

exttt{EMB}}$, which are designed to be more robust against noise and outliers compared to traditional Euclidean distance measures when conducting clustering analyses. (NA?)

  • Consider using a Contrastive Multi-Granularity Learning Framework (CMLF) to effectively extract and fuse multi-granularity temporal information for stock trend prediction tasks, incorporating both cross-granularity and cross-temporal objectives. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

Reinforcement Learning

  • Utilize a reinforcement learning framework to automate the process of prompt engineering for large language models, allowing for end-to-end optimization and improved performance across various downstream tasks. (W. Kong et al. 2024)

  • Focus on developing a comprehensive understanding of the underlying assumptions and limitations of your statistical models, and carefully evaluate the potential impact of these factors on your findings. (Al-Hafez et al. 2023)

  • Utilise online reinforcement learning to align the knowledge of large language models with the environment, thereby improving your ability to solve decision-making problems. (Carta et al. 2023)

  • Utilize the PACE (Prompt with Actor-Critic Editing) methodology to automatically edit and improve the quality of prompts for large language models, leading to increased performance and efficiency. (Yihong Dong et al. 2023)

  • Utilize pretrained large language models (LLMs) to generate diverse, context-sensitive, and human-meaningful goals for exploration in reinforcement learning, thereby improving the efficiency and effectiveness of the learning process. (Yuqing Du, Watkins, et al. 2023)

  • Focus on developing novel prompt-tuning techniques specifically tailored to reinforcement learning (RL) tasks, as opposed to directly applying prompt-tuning approaches from natural language processing (NLP), since RL prompts are more complex and contain environment-specific information. (Shengchao Hu et al. 2023)

  • Utilize a Bayesian safe policy learning framework to ensure that your algorithms maximize the posterior expected value while controlling the posterior expected ACRisk, thus mitigating the risk of producing worse outcomes for specific subgroups. (Z. Jia, Ben-Michael, and Imai 2023)

  • Utilise large language models (LLMs) as a proxy reward function in order to simplify the process of reward design in reinforcement learning (RL) systems. By doing so, users can specify your preferences through natural language prompts, reducing the need for extensive expert demonstrations or complex reward functions. (M. Kwon et al. 2023)

  • Adopt the Direct Preference Optimization (DPO) technique, which allows for direct optimization of a language model to adhere to human preferences without explicit reward modeling or reinforcement learning, thereby simplifying the preference learning process. (Rafailov et al. 2023)

  • Focus on developing query-dependent prompt optimization techniques for large language models, which involves identifying effective prompts for individual queries instead of relying solely on distributional-level prompt optimization. (Hao Sun, Hüyük, and Schaar 2023)

  • Utilise Bayesian Inverse Reinforcement Learning (BIRL) to effectively model the inverse reinforcement learning process. By doing so, they can leverage the power of Bayesian inference to derive a probability distribution over the space of reward functions, thereby enabling them to develop efficient algorithms that find solutions for the reward learning and apprenticeship learning tasks that generalise well over these distributions. (R. Wei et al. 2023)

  • Consider utilizing the Natural Actor-Critic methodology in reinforcement learning tasks, as it offers improved efficiency over traditional approaches through the use of natural policy gradients, which are covariant and require fewer data points for accurate estimation. (R. Zhou et al. 2023)

  • Consider utilizing reinforcement learning techniques in conjunction with deep neural networks to tackle complex natural language processing tasks, particularly in areas such as syntactic parsing, language understanding, text generation, machine translation, and conversational systems. (Uc-Cetina et al. 2022)

  • Utilise large language models (LLMs) for few-shot planning for embodied agents, enabling them to efficiently follow natural language instructions to complete complex tasks in visually-perceived environments. (M. Ahn et al. 2022)

  • Utilise a Bayesian approach to maintaining uncertain information, extending Watkins Q-learning by maintaining and propagating probability distributions over the Q-values, which are then used to compute a myopic approximation to the value of information for each action, thus enabling the selection of the action that best balances exploration and exploitation.’ (F. Che et al. 2022)

  • Consider using hierarchical abstract machines (HAMs) to constrain the policies considered by reinforcement learning algorithms, allowing for the reduction of search spaces and facilitating knowledge transfer across problems and recombination of component solutions for tackling larger, more complex issues. (Furelos-Blanco et al. 2022)

  • Utilise a two-step Bayesian approach to optimise clinical decisions with timing. (Hua et al. 2022)

  • Consider using perturbed MCMC samplers within the ABC and BSL paradigms to significantly accelerate computation while maintaining control over computational efficiency. (Levi and Craiu 2022)

  • Carefully consider the limitations of Markov reward functions in expressing complex tasks, and utilize polynomial-time algorithms to construct suitable reward functions when possible. (Abel et al. 2021)

  • Utilise a robust optimization approach to find an improved policy without inadvertently leading to worse outcomes. This involves partially identifying the expected utility of a policy by calculating all potential values consistent with the observed data, and finding the policy that maximises the expected utility in the worst case. The resultant policy is conservative but has a statistical safety guarantee, allowing the policymaker to limit the probability of yielding a worse outcome than the existing policy. (Ben-Michael et al. 2021)

  • Leverage the coordination graph technique to efficiently compute the optimal joint action in multi-agent systems, reducing the need for extensive communication and observation among agents. (Bouton et al. 2021)

  • Utilise tree-specific effective sample sizes (ESS) to accurately evaluate the mixing and autocorrelation of Markov Chain Monte Carlo (MCMC) samples of phylogenies, thereby enabling better understanding of the Monte Carlo error associated with various phylogenetic quantities. (Magee et al. 2021)

  • Utilise a centralised task dispatching model, an actor-evaluator-learner programming architecture, and a higher-level abstraction of MARL training paradigms when developing a scalable and efficient computing framework for population-based multi-agent reinforcement learning. (M. Zhou et al. 2021)

  • Utilize advanced particle methods and exploit specific aspects of SDEMEMs to improve efficiency and accuracy in parameter inference for stochastic differential equation mixed effects models. (Botha, Kohn, and Drovandi 2021)

  • Employ the PL-Rank method for optimizing PL ranking models, as it significantly reduces computational costs and promotes fairness aspects of ranking models. (Oosterhuis 2021)

  • Consider using Monte Carlo Tree Search for Policy Optimization (MCTSPO) as an alternative to gradient-based methods for policy optimization in deep reinforcement learning, particularly in situations involving deceptive or sparse reward functions. (Grill et al. 2020)

  • Utilise reinforcement learning (RL) as a powerful tool for addressing complex combinatorial optimization problems, leveraging its ability to automatically search for effective heuristics in a supervised or self-supervised manner. (Mazyavkina et al. 2020)

  • Utilize the Policy Pruning and Shrinking (PoPS) algorithm to efficiently train Deep Reinforcement Learning (DRL) models while maintaining strong performance and achieving compact representations of the DNN. (Livne and Cohen 2020)

  • Consider using a History-inspired Navigation Policy (HiNL) framework to effectively estimate navigation states by utilizing historical states, thereby improving the success rate and success weighted by path length in object-goal visual navigation tasks. (W.-Y. Chen et al. 2019)

  • Optimize at the slot-level rather than the slate-level, which makes the approach computationally efficient. (Dimakopoulou, Vlassis, and Jebara 2019)

  • Utilise relational reinforcement learning techniques, which combine Q-learning and logical regression trees, as well as P-learning and logical decision trees, to effectively model and solve problems involving uncertain environments. (Zambaldi et al. 2018)

  • Use a search session Markov decision process (SSMDP) to model multi-step ranking problems in e-commerce applications, allowing for the optimization of long-term accumulative rewards through reinforcement learning techniques. (Yujing Hu et al. 2018)

  • Consider adopting a distributional perspective when working with reinforcement learning models, as it leads to improved performance and stability. (Bellemare, Dabney, and Munos 2017)

  • Optimize your experiment selection strategy in situations where multiple experiments are available and resources are limited, taking into account the opportunity cost of assigning participants to a specific experiment. (Goldberg and Johndrow 2017)

  • Focus on developing scalable, distributed reinforcement learning algorithms that combine decoupled acting and learning with off-policy correction methods like V-trace to achieve stable learning at high throughput, improved data efficiency, and positive transfer between tasks. (Hermann et al. 2017)

  • Utilize hierarchical reinforcement learning (HRL) for dialogue management, specifically through the application of the option framework, as it enables faster learning and superior policy development compared to traditional flat reinforcement learning techniques. (Budzianowski et al. 2017)

  • Utilize deep reinforcement learning to train visual dialog agents end-to-end, from pixels to multi-agent multi-round dialog to game reward, in order to effectively develop goal-driven training for visual question answering and dialog agents. (A. Das et al. 2017)

  • Consider incorporating natural language instructions as a supplementary reward mechanism in reinforcement learning algorithms to enhance your efficiency and effectiveness, especially in environments with sparse rewards. (Kaplan, Sauer, and Sosa 2017)

  • Consider using entropy-regularized reinforcement learning techniques, as they demonstrate a precise equivalence between Q-learning and policy gradient methods in this context, potentially improving the performance and understanding of your models. (Schulman, Chen, and Abbeel 2017)

  • Focus on developing systems that can handle dynamic environments through reinforcement learning, simulated reality, and robust decision-making, while ensuring security and explainability in AI applications. (Stoica et al. 2017)

  • Consider using a Constrained Markov Decision Process (CMDP) framework to optimize bidding strategies in real-time bidding systems, allowing them to balance the need to maximize clicks while staying within budget constraints. (“Advanced Data Mining and Applications” 2017)

  • Consider implementing a distributed and asynchronous version of Guided Policy Search (GPS) to enhance generalization and decrease training times in challenging, real-world manipulation tasks involving multiple robots. (Yahya et al. 2017)

  • Utilise a novel approach to automate feature engineering based on reinforcement learning, which involves training an agent on FE examples to learn an effective strategy of exploring available FE choices under a given budget. (Khurana, Samulowitz, and Turaga 2017)

  • Utilize hierarchical deep reinforcement learning techniques to effectively manage composite tasks, which involve multiple subtasks that must be completed collectively, thereby improving efficiency and user satisfaction. (B. Peng et al. 2017)

  • Consider adopting a reinforcement learning perspective when studying hippocampal function, specifically focusing on the concept of a predictive map’, which represents each state in terms of its ‘successor states’. (Stachenfeld, Botvinick, and Gershman 2016)

  • Formulate the value alignment problem as a cooperative and interactive reward maximization process, specifically through the lens of cooperative inverse reinforcement learning (CIRL), which involves active instruction by the human and active learning by the robot. (Hadfield-Menell et al. 2016)

  • Consider developing a novel learning algorithm called “Reset-free Trial-and-Error” (RTE) that enables complex robots to quickly recover from damage while completing your tasks and taking the environment into account, without requiring a reset to an initial state after each episode. (Pugh, Soros, and Stanley 2016)

  • Utilize a combination of Monte Carlo Tree Search (MCTS) and deep recurrent neural networks (RNN) to efficiently navigate graphs and overcome the challenge of sparse rewards in reinforcement learning tasks. (Bello et al. 2016)

  • Focus on developing a simplified Q-learning algorithm for continuous domains, called normalized advantage functions (NAF), which combines the benefits of policy search and value function estimation without requiring a separate actor or policy function, leading to increased sample efficiency. (S. Gu et al. 2016)

  • Utilize reinforcement learning techniques to develop autonomous optimization algorithms that can adaptively improve your own performance through self-guided policy searches, leading to potentially significant enhancements in convergence speeds and overall objective values compared to traditional hand-engineered algorithms. (Ke Li and Malik 2016)

  • Carefully log propensities and conduct sanity checks to ensure the accuracy of your off-policy learning methods, especially when dealing with large-scale real-world data sets. (Vasile, Lefortier, and Chapelle 2016)

  • Consider incorporating curriculum learning and interactive teaching techniques in your experimental designs to potentially enhance the sample efficiency of grounded language learning systems. (Yonghui Wu et al. 2016)

  • Focus on developing practical algorithms that ensure monotonic improvement through the use of trust regions, which limit the deviation from the original policy during optimization. (Schulman et al. 2015)

  • Utilize the Deep Deterministic Policy Gradient (DDPG) algorithm for continuous control tasks, as it enables end-to-end learning directly from raw pixel inputs, achieving comparable performance to planning algorithms with full knowledge of the domain dynamics. (Lillicrap et al. 2015)

  • Consider the underlying network topology when designing coordination techniques for multiagent systems, as different topologies may significantly affect the coordination performance among agents. (Jianye Hao et al. 2014)

  • Carefully evaluate the performance of various bandit algorithms for tree search, including UCT, Flat-UCB, and BAST, considering factors such as regret bounds, smoothness of rewards, and efficiency in cutting off sub-optimal branches, to determine the most suitable approach for specific applications. (Coquelin and Munos 2014)

  • Apply advanced planning techniques like Upper Confidence Bound in Trees (UCT) to improve the performance of your playlist recommendation systems, particularly in scenarios involving large song libraries. (Xinxi Wang et al. 2013)

  • Carefully consider the type of knowledge to be transferred, the appropriate level of abstraction, and the method of integration when applying transfer learning in multi-agent reinforcement learning domains. (“Recent Advances in Reinforcement Learning” 2012)

  • Carefully consider the implications of policy oscillation and explore the benefits of aggregation-based policy evaluation methods, which offer better error bounds and more regular performance despite having limited cost function representation capabilities. (Bertsekas 2011)

  • Utilise a hierarchical optimistic optimization (HOO) strategy when dealing with X-armed bandit problems, which involves building an estimate of the mean-payoff function f over X, focusing on precision around its maxima while being loose elsewhere, using a binary tree structure to store statistics and guide node selection, and updating the tree based on received rewards. (Bubeck et al. 2010)

  • Consider using a generalized two-filter smoothing formula when working with non-linear non-Gaussian state-space models, as it allows for more flexibility and applicability across different types of models without requiring restrictive assumptions or closed form expressions. (Briers, Doucet, and Maskell 2009)

  • Consider the Bayesian approach to model-based reinforcement learning, which offers an elegant solution to the exploration/exploitation problem by maintaining a distribution over possible models and acting to maximize expected reward, even though the exact computation of the Bayesian policy is often intractable. (Kolter and Ng 2009)

  • Focus on developing accurate heat kernel estimates for jump processes of mixed types on metric measure spaces, taking into account factors like jumping intensities, spatial scales, and temporal dynamics. (Z.-Q. Chen and Kumagai 2007)

  • Carefully consider the assumptions underlying your statistical models, particularly regarding the Markov property, and explore alternative approaches such as reinforced random walks when appropriate. (“Encyclopedia of Biostatistics” 2005)

  • Focus on developing strong solutions to stochastic differential equations involving singular drift terms, particularly in situations where the drift term may not be Lipschitz continuous or dependent on time, and utilizing methods like the Yamada-Watanabe Theorem and the Veretennikov method to establish pathwise uniqueness. (Krylov and Röckner 2004)

  • Utilise variance reduction techniques like control variate methods to improve the accuracy and efficiency of your gradient estimates in reinforcement learning tasks. (P. L. Bartlett, Fischer, and Höffgen 2002)

  • Adopt the Agent Environment Cycle (AEC) model for developing multi-agent reinforcement learning (MARL) applications, as it addresses limitations of previous models and offers advantages such as clearer reward attribution, prevention of race conditions, and closer alignment with how computer games are executed in code. (Bernstein et al. 2002)

  • Consider utilizing the MAXQ method for hierarchical reinforcement learning, which offers advantages such as improved exploration, reduced number of trials required for learning, and faster adaptation to new problems, by leveraging a hierarchical structure that allows for efficient sharing and reuse of subtasks. (Dietterich 1999)

  • Prioritize experience replay in reinforcement learning tasks by focusing on transitions with higher expected learning progress, as measured by the magnitude of your temporal-difference error, to achieve faster learning and better overall performance. (Lecun et al. 1998)

  • Aim for an asymptotically optimal acceptance rate of approximately 0.234 when scaling the proposal distribution of a multidimensional random walk Metropolis algorithm to maximize its efficiency. (A. Gelman, Gilks, and Roberts 1997)

  • Focus on developing algorithms that effectively balance exploration and exploitation in reinforcement learning tasks, while considering various models of optimality such as finite-horizon, infinite-horizon discounted, and average-reward models. (Kaelbling, Littman, and Moore 1996)

  • Utilize Markov Chain Monte Carlo (MCMC) techniques, specifically the Gibbs Sampler, to efficiently explore complex probability surfaces in Bayesian inference, thereby improving the accuracy and reliability of your conclusions. (Besag and Green 1993)

  • Focus on developing algorithms that balance exploration and exploitation in order to optimize decision making under uncertainty, particularly in scenarios involving multiple options with varying potential rewards. (NA?)

  • Consider combining reinforcement learning with other techniques such as experience replay, learning action models for planning, and teaching to accelerate convergence and enhance performance in solving complex learning tasks. (NA?)

  • Carefully consider the choice of algorithmic parameters, scaling issues, and representational strategies when applying temporal difference learning methods like TD(λ) to complex real-world problems. (NA?)

  • Utilise the REINFORCE algorithms for connectionist reinforcement learning, which enable weight adjustments in the direction of the gradient of expected reinforcement without requiring explicit gradient estimation or storage of related information. (NA?)

  • Focus on developing and testing algorithms that can effectively distinguish between gain-optimal and bias-optimal policies in order to achieve optimal performance in cyclical tasks. (NA?)

  • Utilize a constrained optimization problem to minimize the expected cost of a policy while limiting the change in the policy during each update, thus ensuring stability and preventing drastic shifts in behavior. (NA?)

  • Carefully consider the trade-offs between exploration and exploitation in multi-armed bandit problems, focusing on finding near-optimal solutions with high probability using PAC-type bounds, rather than solely optimizing expected cumulative reward. (NA?)

  • Consider utilizing policy gradient reinforcement learning for optimizing complex tasks like quadrupedal locomotion, as demonstrated by the successful application of this methodology in improving the speed of the Sony Aibo robot. (NA?)

  • Consider the apprenticeship learning setting, where a teacher demonstration of the task is available, because it enables achieving near-optimal performance without requiring explicit exploration, making it safer and more efficient for many applications. (NA?)

  • Carefully consider the choice of function approximation method when combining reinforcement learning (RL) and function approximation techniques, as the interaction between them is not well understood and can significantly impact the overall performance of the algorithm. (NA?)

  • Utilise the “Payoff Propagation” algorithm, which is essentially a decision-making equivalent of Belief Propagation in Bayesian Networks, to efficiently compute individual actions that approximately maximise the global payoff function in a collaborative multiagent setting. (NA?)

  • Utilize the UCT algorithm, which combines Monte Carlo planning with bandit theory, to efficiently explore and exploit options in large state-space Markov decision problems, thereby achieving faster convergence to optimal solutions. (NA?)

  • Focus on optimizing the exploration/exploitation tradeoff in discrete Bayesian reinforcement learning using the proposed BEETLE algorithm, which exploits the optimal value functions simple parameterization as the upper envelope of multivariate polynomials.’ (NA?)

  • Adopt a model-free Reinforcement Learning (RL) algorithm called “Delayed Q-learning” because it is the first model-free algorithm proven to be Probably Approximately Correct in Markov Decision Processes (PAC-MDP), making it suitable for efficiently learning optimal policies in unknown environments. (NA?)

  • Focus on developing a framework that translates the problem of maximizing the expected future return exactly into a problem of likelihood maximization in a latent variable mixture model, for arbitrary reward functions and without assuming a fixed time. (NA?)

  • Explore combining offline and online value functions in your UCT algorithm, as doing so can improve the algorithms performance in various ways, including using the offline value function as a default policy during Monte-Carlo simulation, combining the UCT value function with a rapid online estimate of action values, and utilizing the offline value function as prior knowledge in the UCT search tree.’ (NA?)

  • Employ a hierarchical Bayesian approach to multi-task reinforcement learning, allowing for rapid inference of new environments based on previous ones through the use of a strong prior, while simultaneously enabling quick adaptation to unseen environments via a nonparametric model. (NA?)

  • Adopt a unified Bayesian approach to decision-making, integrating concepts from Markovian decision problems, signal detection psychophysics, sequential sampling, and optimal exploration, while considering computational factors such as subjects knowledge of the task and your level of ambition in seeking optimal solutions.’ (NA?)

  • Utilize batch reinforcement learning algorithms in conjunction with multi-layer perceptrons to effectively learn complex behaviors in various domains, such as robot soccer, due to your efficiency in terms of training experience required and ability to handle large and continuous state spaces. (NA?)

  • Aim to minimize free-energy in your study designs, as doing so allows them to better understand both action and perception, replacing traditional optimal policies of control theory with prior expectations about the trajectory of an agents states.’ (NA?)

  • Adopt standardized metrics and benchmarks for empirically evaluating multiobjective reinforcement learning algorithms, enabling reliable comparisons across different algorithms and promoting advancements in the field. (NA?)

  • Consider using the free-energy framework when studying complex systems, as it allows them to optimize a bound on surprise or value while accounting for prior expectations and uncertainty. (NA?)

  • Utilize eligibility traces for off-policy policy evaluation, as it speeds up reinforcement learning, increases robustness against hidden states, provides a connection between Monte Carlo and temporal-difference methods, and allows for greater multiplication of learning through analysis of multiple policies from the same data stream. (NA?)

  • Utilize the FeynRules package to automate the generation of Feynman rules for any Lagrangian, allowing for seamless integration with multiple Monte Carlo event generators, thereby enabling rapid, robust, and flexible analysis of new physics models. (NA?)

  • Utilize a perturbative framework for jet quenching, incorporating both collisional and radiative parton energy loss mechanisms, and implement this into a Monte Carlo event generator like Jewel. (NA?)

  • Consider using universal value function approximators (UVFAs) to improve the efficiency and effectiveness of reinforcement learning systems by enabling better generalization across both states and goals. (NA?)

  • Integrate various fields of study to achieve a comprehensive understanding of information-seeking behavior, considering both extrinsic and intrinsic motivations, and utilizing diverse methodologies such as reinforcement learning, partial observable Markov decision processes, and eye tracking. (NA?)

  • Focus on developing probabilistic, non-parametric Gaussian processes transition models to improve the efficiency of autonomous learning in robotics and control systems, thereby reducing the impact of model errors and enabling faster learning. (NA?)

  • Carefully consider the type of learning policy they employ in your reinforcement learning algorithms, as it significantly impacts the convergence of the algorithm towards optimal policies. (NA?)

  • Focus on developing algorithms that optimize the response surface of a new task instance by selecting policies from a finite library of policies, drawing inspiration from Bayesian optimization techniques to ensure efficiency in the number of policy executions. (NA?)

  • Formulate the bid decision process as a reinforcement learning problem, where the state space is represented by the auction information and the campaigns real-time parameters, and an action is the bid price to set. (NA?)

  • Employ a two-tier optimization process when developing AI agents for complex multi-agent environments, incorporating a population of independent RL agents trained concurrently from thousands of parallel matches, with each agent learning its own internal reward signal and selecting actions using a novel temporally hierarchical representation. (NA?)

  • Consider leveraging simulation-trained neural networks for transferring agile and dynamic motor skills to real-life legged robots, as it offers a cost-effective and efficient solution for developing advanced control policies. (NA?)

  • Utilize the HOO (hierarchical optimistic optimization) algorithm to improve regret bounds in stochastic bandit problems, especially when dealing with complex, high-dimensional data sets. (NA?)

  • Focus on developing PAC style bounds instead of expected regret for the multi-armed bandit problem, as this approach allows for finding a near-optimal arm with high probability within a limited exploration period. (NA?)

Generative Models

  • Consider incorporating fine-grained textual and visual knowledge of key elements in the scene, along with utilizing different denoising experts at different denoising stages, to improve the quality of generated images in text-to-image diffusion models. (Z. Feng et al. 2023)

  • Focus on developing methods that address data scarcity and modeling complexity in order to advance text-to-audio generation. (R. Huang et al. 2023)

  • Investigate how prompt literacy skills develop among EFL students when they engage in an AI-powered vocabulary-image creation project, and whether this development impacts your subsequent vocabulary learning and engagement with generative AI. (Y. Hwang, Lee, and Shin 2023)

  • Utilize discrete state-space diffusion models for controllable layout generation tasks, as they effectively handle structured layout data in discrete representations and learn to progressively infer a noise-free layout from the initial input. (Inoue et al. 2023)

  • Focus on developing continuous latent diffusion models (LDMs) for text-to-audio (TTA) generation, enabling high-quality audio production with improved computational efficiency and allowing for text-conditioned audio manipulations. (Haohe Liu et al. 2023)

  • Carefully evaluate the benefits and drawbacks of various release methods for generative AI systems, taking into account factors like power concentration, social impacts, malicious use, auditability, accountability, and value judgements, and adopt diverse and multidisciplinary perspectives to manage associated risks. (Solaiman 2023)

  • Consider the unique challenges presented by generative AI technologies, including your inherent variability and the need for clear communication of this characteristic to users, when designing applications for human-AI collaboration. (Weisz et al. 2023)

  • Carefully consider the potential benefits of incorporating text-to-image diffusion models into your visual perception tasks, as these models may offer valuable high-level and low-level knowledge that could improve the accuracy and efficiency of your projects. (Wenliang Zhao et al. 2023)

  • Utilize “chained Markov melding” - an extension of traditional Markov melding - to effectively combine chains of Bayesian submodels into a joint model, thereby allowing for accurate integration of multiple, heterogenous datasets. (Manderson and Goudie 2023)

  • Consider using a random inference model when dealing with Variational Autoencoders (VAEs), where the mean and variance functions of the variational posterior distribution are modeled as random Gaussian processes (GPs). This approach can help improve the accuracy of posterior approximation while maintaining the computational efficiency of amortized inference. (Minyoung Kim 2022)

  • Consider using prompt engineering techniques to enhance the effectiveness of your studies involving artificial intelligence, particularly when working with deep generative models. (Dang et al. 2022)

  • Consider utilizing a text-conditioned diffusion model trained on pixel representations of images to generate scalable vector graphics (SVGs) without having access to large datasets of captioned SVGs. (Graikos et al. 2022)

  • Consider using the Latent Shrinkage Position Model (LSPM) for analyzing network data, as it enables automatic inference on the dimensionality of the latent space, reduces computational burden, and retains interpretability. (Gwee, Gormley, and Fop 2022)

  • Focus on improving diffusion models by enhancing your empirical performance or expanding your theoretical capabilities, using a variety of approaches such as denoising diffusion probabilistic models (DDPMs), score-based generative models (SGMs), and stochastic differential equations (Score SDEs), while considering efficient sampling, improved likelihood estimation, and handling data with special structures. (Ling Yang et al. 2022)

  • Consider using an instruction-tuned large language model (LLM) as the text encoder for text-to-audio (TTA) generation, as demonstrated by the significant improvements seen in the proposed Tango models performance compared to previous state-of-the-art models.’ (Yen-Ju Lu et al. 2022)

  • Develop and implement safe latent diffusion (SLD) to effectively remove and suppress inappropriate image parts during the diffusion process, thereby reducing the risk of inappropriate degeneration in diffusion models. (Zehua Sun et al. 2022)

  • Develop machine learning-enabled data-driven models for effective capacity predictions for lithium-ion batteries under different cyclic conditions, specifically by modifying the isotropic squared exponential kernel with an automatic relevance determination structure (Model A) and coupling the Arrhenius law and a polynomial equation into a compositional kernel (Model B) to consider the electrochemical and empirical knowledge of battery degradation. (Kailong Liu et al. 2021)

  • Consider using a diffusion probabilistic model for singing voice synthesis tasks, as it allows for stable training and produces more realistic outputs compared to other approaches such as simple loss or generative adversarial networks. (Jinglin Liu et al. 2021)

  • Carefully select and optimize the tuning parameters for Hamiltonian Monte Carlo kernels within Sequential Monte Carlo samplers to improve the efficiency and accuracy of Bayesian computations. (Buchholz, Chopin, and Jacob 2021)

  • Consider using a data-dependent adaptive prior when working with denoising diffusion probabilistic models (DDPMs) to improve your efficiency and accuracy. (M. Jeong et al. 2021)

  • Consider using a generative flow model for motion style transfer, as it allows for unsupervised learning on unlabelled motion data, efficient inference of latent codes, and the ability to generate multiple plausible stylized motions. (Sverrisson et al. 2020)

  • Adopt a Bayesian workflow approach to modeling disease transmission, utilizing Stans expressive probabilistic programming language and Hamiltonian Monte Carlo sampling for robust, efficient, and transparent model development and inference.’ (Grinsztajn et al. 2020)

  • Utilize JointDistributions, a family of declarative representations of directed graphical models in TensorFlow Probability, to enable various idioms for probabilistic model specification while maintaining a standardized interface to inference algorithms. (Piponi, Moore, and Dillon 2020)

  • Utilize a multi-scale flow architecture based on a Haar wavelet image pyramid when developing a flow-based generative model for molecule to cell image synthesis. This architecture allows for the generation of cell features at different resolutions and scales to high-resolution images, while maintaining the original objective of maximizing the log-likelihood of the data. (Ardizzone et al. 2019)

  • Utilise a comprehensive compilation scheme to convert Stan programs into generative probabilistic programming languages, allowing them to take advantage of the extensive range of existing Stan models for testing, benchmarking, or experimentation with novel features or inference techniques. (Cusumano-Towner et al. 2019)

  • Carefully evaluate and select appropriate methods for scaling Gaussian processes based on factors such as data volume, desired accuracy, and computational resources, considering options like global and local approximations, sparse kernels, and sparse approximations. (Haitao Liu et al. 2018)

  • Consider using variable length Markov chains (VLMCs) instead of traditional high-order Markov chains for analyzing complex systems, as they provide greater flexibility and structural richness, leading to improved prediction accuracy and better understanding of the underlying dynamics. (Sutter 2018)

  • Consider using WaveGrad, a novel conditional generative model for waveform generation that estimates gradients of the data density, as it allows for a flexible tradeoff between inference speed and sample quality, and bridges the gap between non-autoregressive and autoregressive models in terms of audio quality. (Dumoulin et al. 2018)

  • Focus on maximizing the (_{1})-regularized marginal pseudolikelihood of the observed data to efficiently estimate the dependency structure of a generative model without using any labeled training data. (S. H. Bach et al. 2017)

  • Utilise the brms package in R, which enables easy specification of a wide variety of Bayesian single-level and multilevel models, including distributional regression and non-linear relationships, using an intuitive and powerful formula syntax that extends the well-known formula syntax of lme4. (Bürkner 2017)

  • Consider using Snorkel, an end-to-end system for combining weak supervision sources, to rapidly create accurate and diverse training data for machine learning models. (Ratner et al. 2017)

  • Utilize a combination of text-to-image customized data augmentations, content loss for content-style disentanglement, and sparse updating of diffusion time steps to effectively fine-tune pre-trained diffusion models for generating high-quality images in previously unseen styles using minimal data. (Antoniou, Storkey, and Edwards 2017)

  • Utilize deep generative models of vowel inventories to understand the underlying structure of human language, enabling accurate predictions of held-out vowel systems and providing insights into linguistic universals. (Cotterell and Eisner 2017)

  • Use the stick-breaking representation for homogeneous normalized random measures with independent increments (hNRMI) to develop efficient algorithms for slice sampling mixture models, which rely on the derived representation and can be applied to analyze real data. (Favaro et al. 2016)

  • Consider utilizing the Gated PixelCNN model for conditional image generation due to its ability to match or surpass the performance of PixelRNN while being computationally more efficient, allowing for the creation of diverse and realistic images across various contexts. (Abadi et al. 2016)

  • Utilise complex embeddings for link prediction tasks in statistical relational learning, as they offer superior performance compared to traditional methods, particularly in handling antisymmetric relations, while maintaining scalability and simplicity. (Alon, Moran, and Yehudayoff 2015)

  • Utilise Stan, a powerful probabilistic programming language, to perform Bayesian inference and optimization for complex statistical models across various scientific fields. (Andrew Gelman, Lee, and Guo 2015)

  • Use a combination of synchronous and mixed couplings when studying diffusion processes, as they offer better performance than either type alone, especially when dealing with non-constant diffusion matrices or complex systems involving multiple interacting diffusions. (Eberle 2015)

  • Consider utilizing the chain rule to transform a pretrained 2D diffusion model into a 3D generative model for 3D data generation, while addressing the out-of-distribution problem by employing the proposed Perturb-and-Average Scoring technique. (A. X. Chang et al. 2015)

  • Adopt a probabilistic framework for machine learning, which enables accurate representation and management of uncertainty in models and predictions, leading to improved decision-making and optimization. (J. R. Lloyd et al. 2014)

  • Consider using latent Bayesian melding to effectively integrate individual-level and population-level models, leading to improved accuracy in predictions. (Myerscough, Frank, and Leimkuhler 2014)

  • Consider utilizing deep latent Gaussian models (DLGMs) for generating samples from complex distributions, as they offer a flexible framework for modelling hierarchical relationships among variables while maintaining computational efficiency. (Rezende, Mohamed, and Wierstra 2014)

  • Carefully select the optimal parameterization and update grouping strategy for your latent variable models to achieve faster convergence rates and higher-quality results in your analyses. (Asparouhov and Muthén 2014)

  • Utilize automatic differentiation variational inference (ADVI) for scalable and accurate Bayesian inference, particularly in cases involving complex models and large datasets. (Diederik P. Kingma and Welling 2013)

  • Carefully consider the possibility of multiple underlying mechanisms driving event clustering, such as self-excitation, autocorrelation, and external factors, before drawing conclusions about the predominant cause. (Mohler 2013)

  • Consider using a boosting-based conditional density estimation algorithm for solving general problems involving the estimation of the entire distribution of a real-valued label given a description of current conditions, such as in the case of price prediction in auctions. (Boyer and Brorsen 2013)

  • Utilize mixed membership stochastic blockmodels for analyzing complex relational datasets, as these models allow for greater flexibility in handling multi-faceted data points and provide better insights into the underlying structures and dynamics of the system. (Edoardo M. Airoldi, Wang, and Lin 2013)

  • Leverage the inherent tensor structure within the low-order observable moments of latent variable models like Gaussian mixture models, hidden Markov models, and latent Dirichlet allocation to develop computationally and statistically efficient parameter estimation methods. (D. Hsu and Kakade 2012)

  • Utilise formal model-based inference methods that allow for direct estimation of interpretable ecological quantities rather than relying solely on vague suitability indices derived from presence-only data. (Royle et al. 2012)

  • Consider utilizing advanced deep learning techniques, particularly diffusion models, for scaffold hopping tasks in order to achieve higher levels of accuracy and efficiency. (Bickerton et al. 2012)

  • Utilise the DirectLiNGAM approach for estimating causal ordering and connection strengths in linear non-Gaussian structural equation models, as it guarantees convergence to the right solution within a small fixed number of steps if the data strictly adheres to the model. (Kawahara et al. 2010)

  • Focus on developing a deeper understanding of the uses of probability, statistical modeling, and providing good examples when applying the Dempster-Shafer theory. (“Classic Works of the Dempster-Shafer Theory of Belief Functions” 2008)

  • Utilize mixed membership stochastic blockmodels to effectively analyze complex relational datasets, allowing for greater flexibility in understanding the various roles played by individuals within a system. (Edoardo M. Airoldi et al. 2007)

  • Consider using Bayesian Treed Gaussian Process Models to overcome limitations of traditional Gaussian Process Models, such as scalability, stationarity assumptions, and homogeneous predictive errors, in order to improve accuracy and efficiency in nonparametric regression tasks. (Gramacy and Lee 2007)

  • Focus on proving a quenched invariance principle for the paths of the walk, which involves demonstrating that the linear interpolation of the walk, properly scaled, converges weakly to Brownian motion for almost every percolation configuration. (N. Berger and Biskup 2006)

  • Consider using probabilistic modeling approaches when attempting to optimize large scale systems, as these methods offer significant benefits in terms of scalability and adaptability. (“Scalable Optimization via Probabilistic Modeling” 2006)

  • Utilise a fully Bayesian mixture modelling approach, incorporating novel Markov chain Monte Carlo (MCMC) methods like the “reversible jump” sampler, to accurately estimate the number of components and mixture component parameters simultaneously, while providing a richer understanding of the data through the presentation of posterior distributions. (Richardson and Green 1997)

  • Utilize Markov Chain Monte Carlo (MCMC) methods for simulation in complex biostatistical models, allowing them to perform essentially exact Bayesian computations using simulation draws from the posterior distribution. (Andrew Gelman and Rubin 1996)

  • Consider using partially exchangeable random partitions instead of only focusing on exchangeable ones, as they provide a more flexible and robust approach for modeling complex systems. (Pitman 1995)

  • Utilise the Bayesian framework for modelling, which allows them to explicitly state all assumptions using the language of probability theory, thereby enabling them to generate possible datasets and make informed decisions based on the data. (D. M. Wolpert, Ghahramani, and Jordan 1995)

  • Utilize mixtures of Dirichlet processes when dealing with complex statistical models where the closure property of simple Dirichlet processes does not hold. (Kliemann 1987)

  • Focus on developing a tractable approximation to maximum likelihood learning implemented in a layered hierarchical connectionist network, which enables efficient evaluation of complex generative models while avoiding the intractability of considering all possible explanations. (NA?)

  • Consider adopting a discriminative approach to train Markov Logic Networks (MLNs) by optimizing the conditional likelihood of the query predicates given the evidence ones, rather than the joint likelihood of all predicates. (NA?)

  • Utilize nonparametric Bayesian models, specifically those involving Dirichlet processes, to achieve flexible and robust inference while avoiding critical dependence on parametric assumptions. (NA?)

  • Understand the relationship between universal and characteristic kernels in order to effectively use kernel methods in machine learning and pattern analysis. (NA?)

  • Utilize an adaptive algorithm called M-PMC to optimize the performance of importance sampling by iteratively updating both the weights and component parameters of a mixture importance sampling density, thereby improving the accuracy of statistical inferences. (NA?)

  • Utilize Gaussian processes, a non-parametric method for regression, to model instrumental systematics in transmission spectroscopy studies. (NA?)

  • Consider using generative pre-trained transformer (GPT) models for automated compliance checking (ACC) in the Architecture, Engineering, and Construction (AEC) industry, as these models demonstrate promising accuracy rates and do not require additional domain knowledge or term explanation. (NA?)

  • Adopt a Bayesian probabilistic numerical methodology for solving complex numerical problems, allowing them to incorporate prior knowledge and quantify uncertainty in your results. (NA?)

  • Consider utilizing a generative adversarial network (GAN) conditioned with gene expression signatures to effectively design molecules that have a high likelihood of inducing a desired transcriptomic profile, thereby providing an alternative approach to bridge chemistry and biology in the complex field of drug discovery. (NA?)

  • Consider utilizing model-driven engineering (MDE) principles and techniques to enhance the efficiency and effectiveness of prompt engineering processes across various generative AI systems. (NA?)

  • Consider utilizing Normalizing Flows, a type of generative model, for distribution learning because they offer tractable distributions where both sampling and density evaluation can be efficient and exact, addressing limitations found in other generative models like GANs and VAEs. (NA?)

  • Prioritize subject and style keywords in text-to-image generative models, rather than focusing on connecting words or phrasing variations, as these factors do not significantly impact generation quality. (NA?)

  • Consider using a scalable generative model like Chroma for protein design, which offers advantages such as efficient generation of full complexes, sub-quadratic scaling of computation, and flexible sampling capabilities. (NA?)

  • Consider employing a comprehensive theoretical review of the literature on Generative Artificial Intelligence (GAI) to understand its diverse applications and develop new theoretical models for studying GAI in different sectors. (NA?)

Dimensionality Reduction Techniques

  • Consider incorporating fractal parameters, such as the Hurst exponent, into your analyses to improve prediction accuracy and better understand complex phenomena like language. (Alabdulmohsin, Tran, and Dehghani 2024)

  • Utilize the Cholesky decomposition of a correlation matrix to enable effective handling of the positive-definiteness constraint, leading to faster computation of posteriors for selection and shrinkage priors. (R. P. Ghosh, Mallick, and Pourahmadi 2021)

  • Focus on learning the latent structure of data through geodesic estimation, which involves understanding the relationships between data points in a way that accounts for potential measurement errors and noise, ultimately improving the accuracy of downstream analyses. (Madhyastha et al. 2020)

  • Focus on developing anisotropic quantization loss functions that more greatly penalize the parallel component of a datapoints residual relative to its orthogonal component, leading to improved performance in maximum inner product search applications.’ (R. Guo et al. 2019)

  • Focus on developing algorithms that satisfy four crucial properties: being visually accessible, preserving structural integrity, reducing noise, and ensuring robustness. (Moon et al. 2017)

  • Focus on developing algorithms that leverage low-rank spectral decompositions to efficiently solve linear systems, thereby enabling faster and more accurate image retrieval tasks. (Iscen et al. 2017)

  • Consider using anisotropic vector quantization for large-scale inference problems, as it provides significant improvements in accuracy and efficiency compared to traditional quantization methods. (T. Ge et al. 2014)

  • Focus on developing efficient algorithms for performing spectral decomposition and orthogonal matrix factorization, as these techniques can lead to significant improvements in the accuracy and speed of product quantization methods. (Babenko and Lempitsky 2014)

  • Carefully consider the choice of correlation matrix when simulating data for various analyses, as different choices may lead to significantly different results. (Hardin, Garcia, and Golan 2013)

  • Consider using multiple maps t-SNE, an extension of t-SNE, to effectively visualize non-metric similarities in complex datasets, thereby avoiding the limitations imposed by traditional multidimensional scaling methods. (Maaten and Hinton 2011)

  • Use nuclear norm minimization (NNM) to solve affine constrained matrix rank minimization (ACMRM) problems, which involves minimizing the sum of singular values of a matrix subject to certain constraints, because it has been proven to provide accurate solutions under specific conditions. (S. Ma, Goldfarb, and Chen 2009)

  • Use the Singular Value Projection (SVP) algorithm for solving Affine Rank Minimization Problems (ARMP) because it provides a simple, fast, and effective way to recover the minimum rank solution for affine constraints that satisfy the Restricted Isometry Property (RIP), while also offering robustness to noise and improved performance compared to other existing methods. (Meka, Jain, and Dhillon 2009)

  • Focus on studying the singularities of the hypersurface defined by a polynomial to improve the lower bounds for the rank of a symmetric tensor. (Landsberg and Teitler 2009)

  • Utilize a three-way tensor factorization model for collective learning on multi-relational data, as it allows for efficient computation and improved performance compared to existing tensor approaches and state-of-the-art relational learning solutions. (Bader, Harshman, and Kolda 2007)

  • Utilise principal curves - smooth one-dimensional curves passing through the middle of a p-dimensional dataset - as a nonlinear summary tool for understanding complex datasets. (Hastie and Stuetzle 1989)

  • Use the Nyström method to efficiently approximate a Gram matrix for improved kernel-based learning algorithms, which can significantly reduce computational costs while preserving accuracy. (NA?)

  • Focus on developing efficient algorithms for learning similarity-preserving hash functions that map high-dimensional data onto binary codes, while considering scalability and efficiency for large datasets. (NA?)

  • Utilise a new optimisation criterion for discriminant analysis that doesnt require the nonsingularity of the scatter matrices, allowing it to handle undersampled problems effectively. (NA?)

  • Use the Singular Value Decomposition (SVD) to efficiently analyze large datasets, providing a powerful tool for clustering and dimensionality reduction. (NA?)

  • Carefully consider the unique challenges posed by high-dimensional data, including the “curse of dimensionality” and the concentration of norms, and adopt suitable distance measures, kernels, and dimension reduction techniques accordingly. (NA?)

  • Consider using the Generalized Low Rank Approximations of Matrices (GLRAM) algorithm for dimensionality reduction tasks, as it offers a balance between reducing reconstruction errors and maintaining low computation costs, making it suitable for handling high-dimensional data. (NA?)

  • Carefully balance the tradeoff between preserving local distances and dissimilarities during dimensionality reduction, depending on the specific characteristics of your dataset. (NA?)

  • Consider utilizing the Grassmann manifold for subspace-based learning problems, as it provides a unified framework for both feature extraction and classification within the same space, leading to improved performance over traditional methods. (NA?)

  • Consider using Procrustes analysis for manifold alignment, as it enables a mapping that is defined everywhere rather than just on the training data points, while preserving the manifold shape and maintaining the relationship between data points during the alignment process. (NA?)

  • Utilise a combination of nuclear-norm-regularised matrix approximation and maximum-margin matrix factorisation techniques when dealing with matrix completion problems, as this leads to an efficient algorithm for large matrix factorisation and completion that outperforms both individual approaches. (NA?)

  • Utilize sparse canonical correlation analysis (SCCA) to identify the minimum number of features required to maximize the correlation between two sets of variables, thereby improving model interpretability and reducing computational complexity. (NA?)

  • Consider using Transfer Component Analysis (TCA) for domain adaptation tasks, as it enables efficient discovery of a shared latent space underlying multiple domains, thereby reducing the distance between your distributions and allowing for effective cross-domain prediction. (NA?)

  • Consider using multiple kernel learning (MKL) for dimensionality reduction (DR) in order to efficiently analyze high-dimensional data sets, particularly those involving multiple descriptors, thereby enhancing the effectiveness of various applications including object recognition, image clustering, and face recognition. (NA?)

  • Utilise a task-driven dictionary learning approach for your studies, rather than solely focusing on data-driven methods. This involves optimising the dictionary for the specific task at hand, rather than simply aiming for accurate data reconstruction. By doing so, researchers can achieve superior results across a range of tasks including classification, regression, and compressed sensing. (NA?)

  • Combine sparse neighborhood preserving embedding (SNPE) with maximum margin criterion (MMC) methods to create a discriminant sparse neighborhood preserving embedding (DSNPE) algorithm, which effectively integrates Fisher criterion and sparsity criterion for improved face recognition performance. (NA?)

  • Consider using t-SNE, a novel technique for visualizing high-dimensional data, due to its ability to capture both local and global structures effectively, thereby providing clearer insights into complex datasets. (NA?)

  • Consider utilizing low-rank tensor network approximations, distributed tensor networks, and associated learning algorithms to effectively tackle huge-scale optimization problems, thereby converting them into more manageable, smaller, linked, and/or distributed sub-problems. (NA?)

  • Consider the Nystrom method for large-scale kernel learning tasks, especially when there is a large gap in the eigen-spectrum of the kernel matrix, as it can yield a better generalization error bound compared to random Fourier features based approaches. (NA?)

  • Utilise MinHash sketches, a type of randomised summary structure, to perform quick but approximate processing of cardinality and similarity queries on massive data sets. These sketches are mergeable and composable, allowing for addition of elements or union of multiple subsets to be conducted within the sketch space itself. Furthermore, these sketches are a form of locality sensitive hashing (LSH) scheme, making them particularly effective for tasks such as detecting near-duplicate webpages or analys (Broder, n.d.)

  • Utilise spectral properties of your dataset to improve approximation guarantees for the Column Subset Selection Problem (CSSP) and the Nystrom method, particularly for datasets with known rates of singular value decay such as polynomial or exponential decay. (NA?)

  • Focus on developing visualization tools that preserve local and global fidelity, cluster preservation, and outlier identification when interpreting classifiers that output probabilistic predictions. (NA?)

  • Consider utilizing the t-SNE algorithm for visualizing high-dimensional data, as it effectively preserves both local and global structures, reduces the tendency to crowd points together in the center of the map, and outperforms other non-parametric visualization techniques like Sammon mapping, Isomap, and Locally Linear Embedding. (NA?)

  • Utilize data-driven dimension reduction techniques based on transfer operator theory to effectively analyze complex dynamical systems, while being aware of the similarities and differences among various methods like TICA, DMD, and your generalizations. (NA?)

  • Carefully examine the early exaggeration phase of t-SNE embedding in real time to identify optimal conditions for improved visualization of large cytometry datasets. (NA?)

  • Use the proposed Least Squares Linear Discriminant Analysis (LS-LDA) technique for multi-class classifications, as it provides a direct formulation of LDA as a least squares problem, improving its applicability and performance in high-dimensional and undersampled data scenarios. (NA?)

Feature Selection Methods

  • Utilise the Conditional Mutual Information Maximisation (CMIM) criterion for feature selection in classification tasks. This criterion allows for the selection of features that are both individually informative and two-by-two weakly dependent, leading to improved accuracy and reduced overfitting. (A. K. Sinha et al. 2022)

  • Carefully plan your data usage, thoroughly understand your data, consult domain experts, stay updated on advancements in deep learning, and rigorously validate your models through appropriate test sets and statistical tests. (Lones 2021)

  • Consider using a Shapley-value variance decomposition of the familiar R^2 from classical statistics as a model-agnostic approach for assessing feature importance in machine learning prediction models, which fairly allocates the proportion of model-explained variability in the data to each model feature. (Redell 2019)

  • Extend the iteratively sure independent screening (ISIS) method beyond the linear model to a general pseudo-likelihood framework, which includes generalized linear models as a special case, to improve feature selection in high-dimensional spaces. (J. Fan and Lv 2018)

  • Develop a comprehensive understanding of the various aspects involved in feature engineering, such as handling diverse data types, dealing with temporal information, navigating complex relational graphs, and managing large transformation search spaces, in order to effectively automate the process and enhance the overall quality of predictive analytics projects. (Lam et al. 2017)

  • Leverage the training examples mean margins of boosting to select features, using a weight criterion called Margin Fraction (MF) in conjunction with a sequential backward selection method, resulting in a novel algorithm called SBS-MF.’ (Alshawabkeh et al. 2012)

  • Consider using feature hashing for large-scale multitask learning due to its ability to effectively reduce dimensionality and preserve sparsity, leading to improved performance and reduced computational costs. (Weinberger et al. 2009)

  • Utilize the Hilbert-Schmidt Independence Criterion (HSIC) as a measure of dependence between features and labels in supervised feature selection, due to its capability to detect any desired functional dependence and its concentration with respect to the underlying measure. (Le Song et al. 2007)

  • Utilise genetic algorithms as a front-end to traditional rule induction systems in order to optimally select the best subset of features for machine learning tasks, thereby reducing the number of features needed while maintaining high recognition rates even in challenging environments. (NA?)

  • Consider using a fast correlation-based filter method for feature selection in high-dimensional datasets, as it can efficiently identify relevant features and detect redundancies without requiring pairwise correlation analysis. (NA?)

  • Consider using the Hilbert-Schmidt Independence Criterion (HSIC) for feature selection in machine learning applications, as it offers a flexible and effective method for selecting informative feature subsets without requiring explicit density estimation. (NA?)

  • Utilize the Top-Scoring Pair(s)’ (TSP) classifier method for analyzing gene expression profiles from pairwise mRNA comparisons. This method offers advantages such as providing decision rules that involve very few genes and only relative expression values, being both accurate and transparent, offering specific hypotheses for follow-up studies, and being parameter-free, thus avoiding issues like over-fitting and inflated estimates of performance.’ (NA?)

  • Carefully consider the choice of feature selection method and classifier type when working with microarray data, as these choices can greatly impact the accuracy and reliability of the resulting model. (NA?)

  • Pay attention to computational performance metrics like build time and classification speed when choosing machine learning algorithms for implementing in real-world scenarios, as these factors can vary significantly even if the classification accuracy remains high. (NA?)

  • Utilise the mutual information measure to select variables from the initial set in spectrometric nonlinear modelling, as it is model-independent and nonlinear, thereby enabling accurate predictions and maintaining interpretability. (NA?)

  • Employ a Maximal Marginal Relevance (MMR) approach for feature selection in text categorization tasks, as it effectively balances information gain and novelty of information, leading to better performance in comparison to traditional information gain and greedy feature selection methods. (NA?)

  • Carefully consider the choice of appropriate data mining techniques based on the nature of the problem, size of the dataset, and desired outcome, while being mindful of potential limitations and assumptions inherent in those techniques. (NA?)

  • Consider using positive approximation as an effective means to enhance the speed and efficiency of heuristic attribute reduction algorithms in rough set theory without compromising the quality of results. (NA?)

  • Carefully consider the choice between wrapper and filter methods for instance selection, taking into account factors such as computational efficiency, noise tolerance, and the potential impact on classification accuracy. (NA?)

  • Utilize local learning to break down complex nonlinear problems into simpler locally linear ones, allowing for accurate global learning within a large margin framework. (NA?)

  • Consider employing a correlation-based feature selection (CFS) algorithm to improve the efficiency and effectiveness of machine learning algorithms by reducing the dimensionality of the data and allowing learning algorithms to operate faster and more accurately. (NA?)

  • Employ the Iteratively Sure Independent Screening (ISIS) method for feature selection in ultrahigh dimensional spaces, as it extends beyond the limitations of traditional linear models and offers improvements in computational efficiency, statistical accuracy, and algorithmic stability. (NA?)

  • Consider using feature selection methods like Filter, Wrapper, and Embedded techniques to effectively manage high-dimensional data, improve computational efficiency, enhance prediction performance, and gain deeper insights into the underlying processes. (NA?)

  • Carefully consider the trade-off between computational cost and potential overfitting risks when choosing between filter, wrapper, and embedded feature selection methods for analyzing DNA microarray data. (NA?)

  • Consider utilising an ensemble-based multi-filter feature selection method for DDoS detection in cloud computing, which combines the outputs of four filter methods to achieve optimal feature selection, thereby increasing classification accuracy and reducing computational complexity. (NA?)

  • Consider employing dimensionality reduction techniques, specifically feature extraction or feature selection, to overcome the curse of dimensionality in high-dimensional data, thereby improving learning performance, increasing computational efficiency, decreasing memory storage, and building better generalization models. (NA?)

  • Develop more intelligent techniques for selecting an initial set of features from which to start the search, formulate search-control methods that take advantage of structure in the space of feature sets, devise improved frameworks for evaluating the usefulness of alternative feature sets, and design better halting criteria that will improve efficiency without sacrificing useful feature sets. (NA?)

  • Consider using a correlation-based filter algorithm for feature selection in machine learning tasks, as it can improve efficiency and reduce data dimensionality without compromising accuracy. (NA?)

Regularization Techniques

  • Utilise second order methods like Variable Projection (VarPro) to replace non-convex penalties with surrogates that convert the original objectives to differentiable equivalents. This leads to faster convergence rates in comparison to standard splitting schemes like Alternating Direction Methods of Multipliers (ADMM) or other subgradient methods. (Sverrisson et al. 2020)

  • Consider using the oem package for efficient computation of penalized regression models in big tall data scenarios, where the number of observations is much larger than the number of variables, and take advantage of its out-of-memory computation capabilities and optimized cross-validation procedures. (Huling and Qian 2018)

  • Utilize a hierarchical group-lasso regularization technique to learn pairwise interactions in linear regression or logistic regression models, ensuring that whenever an interaction is estimated to be nonzero, both its associated main effects are also included in the model. (M. Lim and Hastie 2015)

  • Utilise the Bayesian bridge estimator for regularised regression and classification tasks, as it offers improved estimation and prediction capabilities, handles sparsity better than alternatives, and leads to an MCMC with superior mixing compared to other heavy-tailed, sparsity-inducing priors commonly used in Bayesian inference. (Polson, Scott, and Windle 2011)

  • Utilise an l1-penalised log-determinant Bregman divergence to estimate the inverse covariance or concentration matrix of a multivariate Gaussian distribution, which corresponds to l1-penalised maximum likelihood in this context. (Ravikumar et al. 2008)

  • Utilize the extended Bayesian Information Criterion (EBIC) for model selection in cases involving large model spaces, as it effectively balances the tradeoff between model fit and complexity, thereby reducing the risk of selecting models with excessively high numbers of spurious variables. (J. Chen and Chen 2008)

  • Utilize penalized discriminant analysis (PDA) to overcome issues arising from large numbers of correlated predictor variables in linear discriminant analysis (LDA) by modifying LDA to effectively regularize a large, nearly or fully degenerate within-class covariance matrix (_{}). (Kliemann 1987)

  • Utilize penalized discriminant analysis (PDA) to overcome issues arising from large numbers of correlated predictor variables in linear discriminant analysis (LDA), particularly in situations where the number-of-variables to sample-size ratio is too high, leading to unreliable covariance matrix estimations. (NA?)

  • Differentiate between class noise and attribute noise when evaluating the impact of noise on machine learning systems, as they have distinct implications for classification accuracy and require separate handling strategies. (NA?)

  • Focus on understanding the choice of the regularization parameter in your least-square regression models, as its proper selection significantly impacts the learning rates and overall model performance. (NA?)

  • Ensure that your loss and penalty functions meet the restricted strong convexity and weak convexity conditions, respectively, to guarantee that any stationary point of the composite objective function lies within statistical precision of the underlying parameter vector. (NA?)

  • Utilize the proposed penalty function for empirical risk minimization procedures to achieve sparse estimators, especially when dealing with situations involving potentially overlapping groups of covariates or a graph of covariates. (NA?)

  • Utilise a cyclical blockwise coordinate descent algorithm when dealing with multi-task Lasso problems, as it enables efficient solving of problems with thousands of features and tasks. (NA?)

  • Adopt a fully Bayesian formulation of the lasso problem, which provides valid standard errors and is based on a geometrically ergodic Markov chain, leading to superior prediction mean squared error performance compared to frequentist lasso methods. (NA?)

Ensemble Methods

  • Use a “feedback-reflect-refine” cycle for prompt ensemble learning, which involves generating new prompts based on the inadequacies of existing ones, thereby reducing potential conflicts and redundancies among prompts and creating a more stable and efficient learner. (Chenrui Zhang et al. 2023)

  • Carefully consider the choice of weights assigned to each expert opinion in logarithmic pooling, as the resulting pooled distribution depends heavily on these weights. (Carvalho et al. 2023)

  • Utilize Bayesian hierarchical stacking to effectively leverage multiple candidate models, allowing for improved model fit and conditional local fit in small and new areas. (Yuling Yao et al. 2022)

  • Use stacking of predictive distributions instead of traditional Bayesian model averaging techniques when dealing with the M-open scenario, where the true data-generating process is not among the candidate models being considered. (Yuling Yao et al. 2018b)

  • Utilize the Mesa framework, which employs a meta-sampler to dynamically adjust the resampling strategy based on the current state of ensemble training, leading to improved performance in imbalanced learning scenarios. (Lu Jiang et al. 2017)

  • Consider implementing a model-parallel online learning algorithm based on decision trees, such as the Vertical Hoeffding Tree (VHT), to achieve parallel, online, highly-accurate classification while maintaining compatibility with any specific online boosting algorithm. (Vasiloudis, Beligianni, and Morales 2017)

  • Focus on developing deep stacked ensembles, which are composed of multiple layers of diverse algorithms and hyperparameter configurations, to achieve superior performance in machine learning tasks. (Wistuba, Schilling, and Schmidt-Thieme 2017)

  • Carefully consider the tradeoff between effectiveness and simplicity when building a promoted listings system, taking into account the current scale of the platform and focusing on optimizing click-through rates (CTR) using various methods such as historical features, content-based features, and ensemble learning. (Aryafar, Guillory, and Hong 2017)

  • Consider utilizing online boosting algorithms, specifically the proposed Online BBM and AdaBoost.OL algorithms, to optimize the accuracy of weak online learning algorithms while accounting for adaptivity and sample complexity constraints. (Beygelzimer, Kale, and Luo 2015)

  • Utilize a novel boosting ensemble method for adaptive mining of data streams, which combines the predictions of multiple base models, each learned using a learning algorithm called the base learner, and extends the traditional boosting technique to handle data streams, thereby enabling faster learning and competitive accuracy using simpler base models. (Díaz et al. 2015)

  • Consider implementing adaptive resampling and combining (ARC) algorithms, specifically the ARC-FS algorithm, when working with unstable classifiers such as decision trees, as it effectively reduces variance and improves classification accuracy without requiring extensive parameter tuning or optimization. (Chandra and Pipil 2013)

  • Extend existing transfer and multitask learning algorithms to operate in an “anytime” setting, allowing for continuous improvement in model performance as additional data becomes available. (Boyu Wang and Pineau 2013)

  • Modify existing boosting algorithms to accommodate the unique characteristics of human learners, such as your limited capacity to process high-dimensional feature vectors and your susceptibility to classification noise, in order to improve the overall performance of human-machine collaborative learning systems. (Grubb and Bagnell 2011)

  • Consider extending the traditional boosting framework by incorporating hidden variables to achieve improved results compared to baseline approaches. (Haffari et al. 2008)

  • Stop the AdaBoost algorithm after n^(1-ε) iterations, where n is the sample size and ε is within the range of (0,1), to ensure that the sequence of risks of the classifiers it produces approaches the Bayes risk. (Reyzin and Schapire 2006)

  • Ensure accurate implementation of the Randomized Maximum Likelihood (RML) method within a Bayesian framework to achieve an adequate representation of the a posteriori distribution for the PUNQ problem, thereby reducing potential bias in predictions. (G. Gao, Zafari, and Reynolds 2005)

  • Use stacked generalization, a technique for combining classifiers, to improve the efficiency of automatically induced anti-spam filters in the field of text categorization. (Sakkis et al. 2001)

  • Adopt the ROC convex hull (rocch) method for evaluating and selecting classifiers in uncertain environments, as it enables identification of potentially optimal classifiers regardless of the specific class and cost distributions. (Provost and Fawcett 2000)

  • Utilise MBoost, a novel extension to AdaBoost, to manage domain knowledge and multiple models simultaneously, thereby providing robustness against overfitting or poor matching of models to data. (Avnimelech and Intrator 1999)

  • Consider implementing adaptive resampling and combining (ARC) algorithms, specifically the ARC-FS algorithm, when working with unstable classifiers such as decision trees, as it effectively reduces variance and improves classification accuracy without requiring extensive parameter tuning or optimization. (NA?)

  • Consider the possibility of transforming a weak learning algorithm into a stronger one through a process of recursive refinement, thereby enhancing the overall performance of the learning system. (NA?)

  • Aim to create diverse and accurate base learners within your ensemble models, as this increases the likelihood of improving overall model performance. (NA?)

  • Carefully consider the choice of combining technique (bagging, boosting, or random subspace method) depending on the specific characteristics of the base classifier and the available training sample size, as each technique has unique strengths and limitations in improving the performance of weak classifiers. (NA?)

  • Utilise the AdaBoost algorithm, a type of boosting methodology, to improve the accuracy of your machine learning models. This involves iteratively selecting and combining multiple weak learners, each trained on a differently weighted version of the original training data, until a stronger overall model is achieved. (NA?)

  • Utilise the AdaBoost algorithm, a powerful machine learning tool, to improve the accuracy of your learning algorithms. It works by iteratively selecting and combining multiple weak learners, each trained on a differently weighted version of the training data, until a strong learner emerges. This process allows the algorithm to focus on the hardest examples in the training set, thereby increasing overall prediction accuracy. (NA?)

  • Consider using ensemble selection techniques to improve the performance of your models, particularly when dealing with large datasets and various performance metrics. (NA?)

  • Carefully evaluate and optimize the trade-off between diversity and accuracy when selecting a set of base classifiers for your ensemble learning algorithm, considering factors like the cost function being optimized and the potential need for sacrificing some base classifier accuracy to achieve greater overall ensemble diversity. (NA?)

  • Consider using the AdaBoost algorithm for network intrusion detection due to its ability to effectively handle diverse feature types, reduce overfitting, and maintain low computational complexity while achieving high detection rates and low false-alarm rates. (NA?)

  • Incorporate confidence-weighted linear classifiers into your models, which adds parameter confidence information to linear classifiers and enables online learners to update both classifier parameters and the estimate of your confidence. (NA?)

  • Consider using ensemble methods, particularly AdaBoost, for improving the performance of weak learners in classification tasks, as they can generate a final classifier with reduced misclassification rate and lower variance compared to the base learner. (NA?)

  • Consider the five dimensions of ensemble methods in classification tasks: inducer, combiner, diversity, size, and members dependency, along with selection criteria from the practitioners perspective, to choose the most appropriate ensemble method for your specific application.’ (NA?)

  • Consider using the SemiBoost algorithm, a boosting framework for semi-supervised learning, to improve the classification accuracy of any given supervised learning algorithm by leveraging available unlabeled examples. (NA?)

  • Focus on creating diverse and accurate classifiers to improve the overall performance of ensemble methods in machine learning. (NA?)

  • Combine all available imaging modalities together in a single automated learning framework, allowing for a clearer view of the progression of disease pathology. (NA?)

  • Utilize diverse ensemble methods to effectively manage concept drift in online learning systems, as this approach leads to superior performance compared to traditional methods. (NA?)

  • Consider utilizing an ensemble of detectors and background knowledge to effectively label events in unlabeled data, particularly when human expertise is unavailable or impractical. (NA?)

Transfer Learning

  • Consider implementing multitask prompt tuning (MPT) for efficient transfer learning, which involves learning a single transferable prompt by distilling knowledge from multiple task-specific source prompts, followed by applying multiplicative low rank updates to adapt it to each downstream target task. (Zhen Wang et al. 2023)

  • Develop a deep understanding of the underlying causes of endogenous shifts in cross-domain detection tasks, and then use techniques such as local prototype alignment and global adversarial learning to effectively suppress those perturbations. (Tao et al. 2022)

  • Consider applying computational intelligence techniques like neural networks, Bayesian networks, and fuzzy logic to enhance the efficiency and accuracy of transfer learning methods. (Zamini and Kim 2022)

  • Utilize a multi-task adaptive Bayesian linear regression model for transfer learning in Bayesian optimization, as it enables efficient sharing of information across related black-box optimization problems and leads to significant improvements in speed and accuracy. (Yang Li et al. 2022)

  • Consider using off-the-shelf inertial measurement unit (IMU) datasets as the source domain for building activity recognition models for millimeter wave (mmWave) radar sensors, allowing for more efficient deployment and reducing the need for extensive in-situ data collection and labeling costs. (Bhalla, Goel, and Khurana 2021)

  • Utilize the Wasserstein Barycenter Transport (WBT) method for multi-source domain adaptation, which involves creating an intermediate domain between multiple source domains and the target domain using the Wasserstein barycenter, followed by transporting the sources to the target domain using standard Optimal Transport for Domain Adaptation framework. (Turrisi et al. 2020)

  • Utilise stabilised regression when dealing with multi-environment regression scenarios, as it enables them to identify stable and unstable predictors, thereby improving generalisation performance to previously unseen environments. (Pfister et al. 2019)

  • Consider using a mixture-of-experts approach for unsupervised domain adaptation from multiple sources, which involves explicitly capturing the relationship between a target example and different source domains using a point-to-set metric, and learning this metric in an unsupervised fashion using meta-training. (Jiang Guo, Shah, and Barzilay 2018)

  • Consider using a Slimmable Domain Adaptation approach to improve cross-domain generalization while allowing for architecture adaptation across various devices. (Brock et al. 2017)

  • Focus on aligning infinite-dimensional covariance matrices in reproducing kernel Hilbert spaces (RKHS) for effective domain adaptation, rather than solely focusing on reducing distribution discrepancies in input spaces. (Courty et al. 2017)

  • Consider whether your data allows for label-preserving transformations, and if so, they should prioritize data augmentation in data-space rather than feature-space for optimal performance in machine learning classification tasks. (S. C. Wong et al. 2016)

  • Consider using a learnable similarity function as the fundamental component of clustering, allowing for successful cross-task and cross-domain transfer learning. (Amid, Gionis, and Ukkonen 2016)

  • Utilise a broad class of ERM-based linear algorithms that can be instantiated with any non-negative smooth loss function and any strongly convex regulariser, as this allows for generalisation and excess risk bounds to be established, leading to improved learning rates. (Kuzborskij and Orabona 2016)

  • Consider organizing your transfer learning schemes carefully to optimize results, taking into account factors such as whether to use consecutive transfer schemes, the similarity of datasets/tasks involved, and the degree of fine-tuning applied. (Menegola et al. 2016)

  • Use Domain Consensus Clustering (DCC) to better exploit the intrinsic structure of the target domain when dealing with Universal Domain Adaptation (UniDA) problems, separating common classes from private ones and differentiating private classes themselves. (G. Hinton, Vinyals, and Dean 2015)

  • Optimize your statistical models by considering both the discriminativeness and domain-invariance of your features, which can be achieved by jointly optimizing the underlying features along with two discriminative classifiers - the label predictor and the domain classifier. (Ganin and Lempitsky 2014)

  • Consider adopting Universal Domain Adaptation (UDA) as a more practical approach to domain adaptation, which involves identifying and adapting to the common label set between source and target domains without assuming prior knowledge about the target domain label set. (Tzeng et al. 2014)

  • Consider using the proposed masked optimal transport (MOT) methodology for partial domain adaptation, as it addresses the limitations of traditional optimal transport (OT) approaches through a combination of relaxation and reweighting techniques, while maintaining theoretical equivalence to conditional OT. (“Inaugural Image and Vision Computing Outstanding Young Researcher Award Winner Announced” 2012)

  • Utilise a feature-level domain adaptation’ (FLDA) approach when dealing with domain adaptation issues in machine learning. FLDA involves modelling the dependence between the source and target domains using a feature-level transfer model, which is then used to train a domain-adapted classifier. This approach is particularly useful when the transfer can be naturally modelled via a dropout distribution, allowing the classifier to adapt to differences in the marginal probability of features in the source (Geoffrey E. Hinton et al. 2012)

  • Not treat instances within a bag as independently and identically distributed (i.i.d.) samples, but rather explore relationships among instances to improve the performance of multi-instance learning models. (Z.-H. Zhou, Sun, and Li 2008)

  • Consider using a Bayesian undirected graphical model for co-training, which provides a principled approach for semi-supervised multi-view learning, clarifying assumptions and offering improvements over traditional co-regularization techniques. (Blum and Mitchell 1998)

  • Leverage recent advances in machine learning to develop efficient approximations for semi-supervised learning that are linear in the number of images, allowing for effective analysis of massive image collections. (NA?)

  • Carefully consider the choice of alpha when combining source and target error in domain adaptation, as the optimal alpha depends on factors such as the divergence between the domains, the sample sizes of both domains, and the complexity of the hypothesis class. (NA?)

  • Consider incorporating a data-dependent regularizer based on the smoothness assumption into your least-squares support vector machines (LS-SVM) models to ensure that the target classifier shares similar decision values with the auxiliary classifiers from relevant source domains on the unlabeled patterns of the target domain. (NA?)

  • Consider utilizing multi-model knowledge transfer techniques to effectively leverage prior knowledge when learning object categories from limited samples, thereby improving the accuracy and efficiency of the learning process. (NA?)

  • Carefully consider what knowledge to transfer, how to transfer it, and when to transfer it in order to effectively utilize transfer learning techniques for improved performance in target domains. (NA?)

  • Consider using a domain-dependent regularizer based on smoothness assumption to ensure that the target classifier shares similar decision values with the relevant base classifiers on the unlabeled instances from the target domain, thereby improving the accuracy of domain adaptation. (NA?)

  • Consider utilising Domain Adaptation Extreme Learning Machines (DAELM) for handling sensor drift issues in e-nose systems. (NA?)

  • Carefully consider the degree of similarity between your source and target domains when applying transfer learning techniques, as well as the type of information transfer (instances, features, parameters, or relationships) that would be most appropriate for your specific situation. (NA?)

  • Carefully choose an appropriate heterogeneous transfer learning (HTL) method based on the availability of labels in your target task, considering factors like the number of target labels, the presence of unlabeled target instances, and the requirement of source labels. (NA?)

  • Carefully consider the type of domain adaptation approach they adopt when dealing with cross-domain generalization problems, taking into account factors such as sample-based, feature-based, and inference-based methods, as well as the assumptions required for performance guarantees. (NA?)

  • Carefully consider the compatibility of source and target tasks in Transfer Learning (TL) to ensure positive transfer and prevent negative transfer, which can lead to reduced performance in the target task. (NA?)

Active Learning

  • Utilise a novel Bayesian method for optimal experimental design by sequentially selecting interventions that minimize the expected posterior entropy as quickly as possible. (Zemplenyi and Miller 2023)

  • Employ active learning algorithms to strategically select experiments that maximize the information gained about the underlying causal structure, thus reducing the overall number of observations needed to accurately infer the structure. (Ben-David and Sabato 2021)

  • Focus on deriving non-trivial general-purpose bounds on label complexity in the agnostic PAC model, specifically by analyzing the performance of algorithms such as \(A^2\) in terms of your dependence on the disagreement coefficient, which measures the growth rate of the region of disagreement as a function of the radius of the version space. (D. J. Foster et al. 2021)

  • Consider implementing active learning techniques, particularly in situations involving imbalanced classes or high similarity among documents, as it can significantly reduce the cost of labeling data and improve the efficiency of supervised learning. (Ducoffe and Precioso 2015)

  • Consider using the Reducible Holdout Loss Selection (RHO-LOSS) method for selecting data points during training, as it effectively filters out less useful samples, improves model performance, and speeds up training across various datasets, modalities, architectures, and hyperparameter choices. (Alain et al. 2015)

  • Consider implementing the (A^{2}) algorithm, which is an agnostic active learning approach that achieves exponential improvement in sample complexity compared to traditional supervised learning methods, particularly in cases involving arbitrary forms of noise. (Beygelzimer et al. 2010)

  • Consider using uncertainty sampling, a sequential approach to sampling, which involves iteratively labelling examples, fitting a classifier from those examples, and using the classifier to select new examples whose class membership is unclear, leading to significant reductions in the number of examples needed to be labelled to produce a classifier with a desired level of effectiveness. (Lewis and Gale 1994)

  • Utilise a novel approach to active learning that specifically designs batches of new training examples and enforces them to be diverse with respect to your angles. (NA?)

  • Use the Agnostic Active Learning (A^2) algorithm to optimize your hypothesis selection process in machine learning tasks, particularly when dealing with noisy or uncertain data. (NA?)

  • Adopt a transductive experimental design approach for active learning, which involves selecting data points that are both hard-to-predict and representative of unexplored test data, leading to improved scalability compared to traditional experimental design methods. (NA?)

  • Focus on developing a deep understanding of label complexity, including the quantities upon which it depends, in order to fully exploit the potential benefits of active learning. (NA?)

  • Utilize the SUMO Toolbox, a comprehensive, adaptive machine learning toolkit, to construct accurate surrogate models for complex systems while minimizing computational costs and maximizing model accuracy. (NA?)

  • Utilize the Free Energy Principle to optimize your experimental designs, as it provides a framework for understanding how organisms interact with your environments and make decisions based on minimizing surprise. (NA?)

  • Consider implementing an active learning approach to the fitting of machine learning interatomic potentials, specifically utilizing the D-optimality criterion for selecting atomic configurations on which the potential is fitted. (NA?)

  • Utilise committee-based sample selection techniques to efficiently train probabilistic classifiers, thereby significantly reducing annotation costs without compromising performance. (NA?)

Neural Networks And Deep Learning

  • Consider developing hardware accelerators based on silicon photonics to improve the performance and energy efficiency of large language models and graph neural networks, as these accelerators offer significant advantages over traditional electronic hardware accelerators. (Afifi et al. 2024)

  • Focus on developing specialized models tailored to individual prompts, rather than attempting to create generalized models capable of handling multiple prompts. (Arar et al. 2024)

  • Carefully consider the potential impact of data contamination on language model performance, specifically focusing on both text and ground-truth contamination, and conduct thorough contamination assessments using appropriate definitions and techniques. (M. Jiang et al. 2024)

  • Consider developing a training approach that allows prompts to extract rich contextual knowledge from LLM data when adapting CLIP for downstream tasks, enabling zero-shot transfer of prompts to new classes and datasets. (Khattak et al. 2024)

  • Consider leveraging the power of pre-trained models, such as Imagebind, to enable effective cross-modal alignment and transfer of knowledge across different domains, ultimately leading to improved performance in tasks such as passive underwater vessel audio classification. (Zeyu Li et al. 2024)

  • Use a combination of 3D molecule-text alignment and 3D molecule-centric instruction tuning to enable language models to better interpret and analyze 3D molecular structures. (Sihang Li et al. 2024)

  • Conduct user studies involving real students to evaluate the efficacy of large language models (LLMs) in computing education, as opposed to merely evaluating LLM outputs through expert review. (Prather et al. 2024)

  • Consider adapting your image-based vision-language models to video through a two-stage process: first, fine-tuning the visual encoder while freezing the language component, and then fine-tuning the language encoder while freezing the visual component. This allows for better utilization of limited video-text data and preserves the diverse capabilities of the original language decoder. (Yue Zhao et al. 2024)

  • Prioritize developing and optimizing prompt strategies for large language models (LLMs) in order to maximize your effectiveness in log analysis tasks, ultimately leading to improved interpretability and adaptability in online scenarios. (“2023 IEEE/ACM 31st International Symposium on Quality of Service (IWQoS)” 2023)

  • Consider incorporating a fully Bayesian Variational Information Bottleneck (BVIB) framework into your statistical shape modeling (SSM) studies, as it allows for the direct prediction of probabilistic anatomy shapes from images while accounting for both aleatoric and epistemic uncertainty. (J. Adams and Elhabian 2023)

  • Pay close attention to the scaling laws governing mixed-modal generative language models, as they capture the complex interactions between individual modalities and help optimize model performance. (Aghajanyan et al. 2023)

  • Carefully consider the choice of prompting strategies when evaluating the performance of generative AI models in multilingual settings, as different approaches may lead to significant differences in performance, particularly for low-resource languages. (Ahuja et al. 2023)

  • Utilise classical PAC-Bayes bounds when analysing the performance of prompted vision-language models, as these bounds offer remarkably tight explanations for the observed performance, even in large domains. (Akinwande et al. 2023)

  • Consider employing a two-branch prompt-tuning paradigm when working with large pre-trained visual-language models (VLMs) for unsupervised domain adaptation (UDA) tasks. The base branch would focus on integrating class-related representation into prompts, ensuring discrimination among different classes, while the alignment branch would utilise image-guided feature tuning (IFT) to make the input attend to feature banks, effectively integrating self- (S. Bai et al. 2023)

  • Consider integrating large pretrained vision-language models directly into low-level robotic control systems to enhance generalization and enable emergent semantic reasoning capabilities. (Brohan et al. 2023)

  • Integrate computational creativity evaluation methodologies into your study designs to effectively analyze and compare the performance of different generative deep learning models in terms of creativity, while considering the potential benefits and drawbacks of various approaches. (M. Chang et al. 2023)

  • Consider using QLoRA, an efficient fine-tuning approach that reduces memory usage while maintaining full 16-bit finetuning task performance, enabling the fine-tuning of larger models on limited hardware resources. (Dettmers et al. 2023)

  • Combine vision-language models (VLMs) and text-to-video models to create a video language planning (VLP) algorithm that allows for efficient and effective long-horizon planning in complex tasks involving both high-level semantics and low-level dynamics. (Yilun Du et al. 2023)

  • Use the proposed SparseGPT algorithm for efficient and accurate pruning of large-scale generative pretrained transformer (GPT) family models, allowing for significant reductions in model size and computational requirements without compromising performance. (Frantar and Alistarh 2023)

  • Consider employing a combination of prefix-tuning and adapter techniques, specifically through an early fusion strategy and bias tuning, to create a parameter-efficient visual instruction model that can effectively handle multi-modal instruction-following tasks. (P. Gao et al. 2023)

  • Focus on developing a comprehensive understanding of the relationship between the number of neurons, the learning rate, and the initialization method in order to effectively train a two-layer neural network with exponential activation functions. (Yeqi Gao, Song, and Yin 2023)

  • Incorporate a chain of thought prompt tuning for vision-language models to achieve improved generalizability, transferability, and domain adaptation across various tasks such as image classification, image-text retrieval, and visual question answering. (J. Ge et al. 2023)

  • Conduct a comprehensive survey of cutting-edge research in prompt engineering on three types of vision-language models: multimodal-to-text generation models, image-text matching models, and text-to-image generation models, focusing on prompting methods, applications, and responsible AI considerations. (J. Gu et al. 2023)

  • Consider using a compact parameter space for diffusion fine-tuning, specifically focusing on singular value decomposition of weight kernels, to achieve better efficiency and effectiveness in personalizing and customizing large-scale text-to-image diffusion models. (L. Han et al. 2023)

  • Utilise a multi-task learning approach when dealing with heterogeneous fashion tasks, which allows for significant improvements in parameter efficiency and model performance compared to traditional single-task models. (X. Han et al. 2023)

  • Consider using Re-parameterized Low-rank Prompts (RLP) for efficient and effective adaptation of vision-language models, particularly in resource-constrained situations. This approach reduces the number of tunable parameters and storage space required, while maintaining or improving performance compared to state-of-the-art methods. (T. Hao et al. 2023)

  • Consider the potential for political bias in conversational AI systems, particularly those designed to provide guidance on political issues, and ensure they account for this in your experimental designs. (Hartmann, Schwenzow, and Witte 2023)

  • Utilize the MGTBench framework to effectively compare and evaluate various machine-generated text detection methods against powerful large language models like ChatGPT-turbo and Claude, considering factors such as transferability, adaptation, and robustness to adversarial attacks. (Xinlei He et al. 2023)

  • Consider incorporating 3D spatial information into large language models through the use of 3D feature extraction and localization mechanisms, enabling the models to better capture and reason about complex 3D scenarios. (Hong et al. 2023)

  • Focus on developing efficient mechanisms like Distilling step-by-step’, which effectively leverages the reasoning capabilities of large language models (LLMs) to train smaller, task-specific models with reduced training data and model sizes, thereby addressing the challenge of deploying LLMs in practical applications.’ (C.-Y. Hsieh et al. 2023)

  • Focus on achieving a balance between model accuracy and complexity when developing algorithms for class incremental learning (CIL), specifically by introducing dense connections between intermediate layers of task expert networks to facilitate knowledge transfer and reduce model growth rates. (Zhiyuan Hu et al. 2023)

  • Utilize a dual-alignment strategy when developing prompts for vision-language models. This involves aligning the prompts with both the knowledge of a large language model (LLM) and local image features. This approach allows the model to benefit from both the implicit context modeling of learnable prompts and the explicit context descriptions provided by the LLM, leading to improved performance on downstream tasks. (Hongyu Hu et al. 2023)

  • Use Scaled Prompt-Tuning (SPT) for few-shot natural language generation tasks because it significantly outperforms traditional Prompt-Tuning with minimal additional training cost, demonstrating improved transferability and offering a solution for data-deficient and computationally limited situations. (T. Hu, Meinel, and Yang 2023)

  • Carefully consider the unique characteristics of point-cloud data and point-based neural network architectures when extending successful 2D channel pruning techniques to 3D point-based networks, rather than simply applying these techniques directly. (Yaomin Huang et al. 2023)

  • Consider incorporating explicit geometry clues into your networks to improve feature learning and downsampling processes, as demonstrated by the successful implementation of the GeoSpark plug-in module. (Zhening Huang et al. 2023)

  • Expand your scope of investigation beyond gender and racial bias in vision-language models to include other relevant groups such as those based on religion, nationality, sexual orientation, or disabilities, and develop appropriate benchmarks for these groups to facilitate comprehensive bias assessments. (Janghorbani and Melo 2023)

  • Develop methods that actively decide when and what to retrieve throughout the generation process, rather than relying on passive retrieval strategies or fixed intervals. (Zhengbao Jiang et al. 2023)

  • Consider using a pre-trained text-to-image diffusion model like Stable Diffusion, and modifying it with motion dynamics and cross-frame attention to create temporally consistent video generation without the need for extensive training or optimization. (Khachatryan et al. 2023)

  • Focus on developing a watermarking technique for large language models that can be efficiently detected without requiring access to the model parameters or API, ensuring that the watermark remains intact even when only a portion of the generated text is used, and providing a rigorous statistical measure of confidence in the detection of the watermark. (Kirchenbauer et al. 2023)

  • Utilize vectorized training to optimize multiple object models simultaneously, thereby improving optimization speed and allowing for efficient handling of large numbers of objects. (X. Kong et al. 2023)

  • Utilise the newly introduced AIOZ-GDANCE dataset to investigate group dance generation, rather than solely focusing on single-dancer choreography. (N. Le et al. 2023)

  • Consider using equivariant shape representations and a novel expectation maximization algorithm to improve unsupervised 3D object segmentation in complex scenes. (Lei et al. 2023)

  • Carefully consider the influence of visual instructions on object hallucination in large vision-language models, as objects that frequently appear in the visual instructions or co-occur with the image objects are more prone to be hallucinated. (Bo Li, Fang, et al. 2023)

  • Evaluate ChatGPTs performance across seven fine-grained information extraction tasks, considering metrics such as performance, explainability, calibration, and faithfulness, to gain a comprehensive understanding of its capabilities.’ (Bo Li, Fang, et al. 2023)

  • Consider utilizing a two-stage pre-training approach when working with large language models and frozen image encoders, specifically focusing on vision-language representation learning followed by vision-to-language generative learning, to improve efficiency and effectiveness in vision-language tasks. (Junnan Li et al. 2023)

  • Consider using a prompt-driven 3D medical image segmentation model like ProMISe, which leverages knowledge from a pretrained 2D image foundation model and integrates lightweight adapters to extract depth-related spatial context without updating the pretrained weights, leading to superior performance compared to state-of-the-art segmentation methods. (Hao Li et al. 2023)

  • Integrate the benefits of existing methods to create a training-efficient method for temporal-sensitive Video Foundation Models (VFMs) that increases data efficiency and enables faster convergence and multimodal friendliness. (Kunchang Li et al. 2023)

  • Avoid narrowly evaluating sparse neural networks (SNNs) on a single or a few tasks and well-understood datasets, and instead use a diverse and challenging benchmark like “Sparsity May Cry” (SMC-Bench) to ensure a comprehensive assessment of SOTA sparse algorithms. (Shiwei Liu et al. 2023)

  • Develop more sophisticated benchmarks in textual inference to improve NLU systems logical reasoning abilities further.’ (Hanmeng Liu et al. 2023)

  • Consider integrating multiple modalities (such as graph, image, and text) in molecular science projects, as doing so can lead to improved accuracy and flexibility in tasks such as molecule generation, molecule captioning, molecular image recognition, and molecular property prediction. (Pengfei Liu et al. 2023)

  • Prioritize developing and optimizing prompt strategies for large language models (LLMs) in order to maximize your effectiveness in log analysis tasks, ultimately leading to improved interpretability and adaptability in online scenarios. (Yilun Liu et al. 2023)

  • Consider employing a mixed scale feature pyramid when dealing with scale variations in object detection tasks, as it allows for improved pseudo label generation and scale-invariant learning. (L. Liu et al. 2023)

  • Consider employing a two-stage pipeline architecture when dealing with imbalanced datasets, particularly in the context of detecting self-stimulatory behaviors in children. (Lokegaonkar et al. 2023)

  • Consider using Error Analysis Prompting (EAPrompt) combined with Chain-of-Thoughts (CoT) and Error Analysis (EA) to enable large language models like ChatGPT to provide human-like translation evaluations at both the system and segment levels. (Q. Lu et al. 2023)

  • Carefully monitor and assess the potential for catastrophic forgetting in large language models during continual fine-tuning, as it can lead to significant loss of previously learned information and negatively impact overall model performance. (Y. Luo et al. 2023)

  • Adopt the Faithful CoT framework, which ensures the reasoning chain provides a faithful explanation of the final answer through a two-stage process of translation and problem solving, thereby enhancing interpretability and improving empirical performance. (Q. Lyu et al. 2023)

  • Consider employing a novel diffusion transformer architecture called DiT-3D for 3D shape generation, which effectively performs denoising operations on voxelized point clouds, leading to improved performance and scalability. (Mo et al. 2023)

  • Utilise a decomposition pipeline when teaching Transformer Language Models to perform arithmetic operations, as it significantly increases your accuracy and effectiveness. (Muffo, Cocco, and Bertino 2023)

  • Utilise Instance-aware Farthest Point Sampling (IA-FPS) and Box-aware Dynamic Convolution to improve the efficiency and accuracy of 3D instance segmentation tasks. (Ngo, Hua, and Nguyen 2023)

  • Focus on developing latent flow diffusion models (LFDM) for conditional image-to-video generation, which involves synthesizing a temporally-coherent flow sequence in the latent space based on the given condition to warp the given image. (Ni et al. 2023)

  • Carefully consider the choice of pre-trained models for specific software engineering tasks, taking into account factors such as architecture, modality, pre-training tasks, and programming languages, as these choices can significantly affect the performance of the models. (C. Niu et al. 2023)

  • Aim to develop data attribution methods that balance computational efficiency and effectiveness, particularly in large-scale, non-convex settings like deep neural networks. (S. M. Park et al. 2023)

  • Consider using modular deep learning techniques to improve the performance, scalability, and robustness of your machine learning models, particularly in situations involving multiple tasks, domain adaptation, and transfer learning. (Pfeiffer et al. 2023)

  • Consider using Imitation learning from Language Feedback (ILF) as a novel approach to improve the alignment of pretrained language models with human preferences, leveraging richer language feedback rather than relying solely on comparison feedback. (Scheurer et al. 2023)

  • Focus on creating a stored instruction computer that connects a language model to an associative memory, following a simple instruction cycle where the next input prompt to be passed to the language model is retrieved from memory, the output of the language model is parsed to recover any variable assignments that are then stored in the associative memory, and the next instruction is retrieved. This approach enables the simulation of a universal Turing machine without modifying the language model weights, thus expanding the range of computations that can (Schuurmans 2023)

  • Consider employing FlexGen, a high-throughput generation engine designed specifically for running large language models (LLMs) with limited GPU memory, which enables efficient patterns to store and access tensors, compresses weights and attention caches, and increases maximum throughput. (Sheng et al. 2023)

  • Combine neural network-based methods with symbolic knowledge-based approaches to develop more capable and flexible AI systems that can address both algorithm-level (abstraction, analogy, reasoning) and application-level (explainable and safety-constrained decision-making) needs. (Sheth, Roy, and Gaur 2023)

  • Consider employing more sophisticated off-the-shelf optimization methods such as Limited memory BFGS (L-BFGS) and Conjugate gradient (CG) with line search instead of stochastic gradient descent methods (SGDs) for deep learning tasks, as these methods can significantly simplify and speed up the process of pretraining deep algorithms. (Shulman 2023)

  • Consider leveraging the interactive capabilities of large-scale language models like ChatGPT to improve the accuracy and efficiency of automated program repair processes. (Sobania et al. 2023)

  • Consider using variational inference to optimize jointly the prompts in a two-layer deep language network (DLN-2), allowing for improved performance compared to a single layer. (Sordoni et al. 2023)

  • Consider implementing Visual Prompt Adaptation (VPA) as a fully test-time and storage-efficient adaptation framework that uses both additive and prependitive adaptable tokens to improve the robustness of vision models. (Jiachen Sun et al. 2023)

  • Consider using the AutoHint framework to improve the efficiency and effectiveness of your large language model (LLM) applications by optimizing prompts through automated hint generation, thereby combining the benefits of both zero-shot and few-shot learning. (Hong Sun et al. 2023)

  • Consider combining prompt tuning and parameter-efficient networks for efficient vision-language model adaptation, particularly in cases where data availability is limited. (Jingchen Sun et al. 2023)

  • Adopt a modular approach to developing complex visual reasoning systems, combining pre-existing models and modules in a sequential manner, guided by a high-level program generated by a large language model. (Surís, Menon, and Vondrick 2023)

  • Consider using the Trainable Projected Gradient Method (TPGM) for fine-tuning pre-trained models, as it allows for automatic learning of distance constraints for each layer, leading to improved out-of-distribution (OOD) performance while retaining generalization capability. (J. Tian et al. 2023)

  • Leverage visual attributes to improve the robustness of transfer learning in Vision-Language (V&L) models, specifically by implementing Attribute-Guided Prompt Tuning (ArGue) to better understand correct rationales and reduce reliance on spurious correlations. (X. Tian et al. 2023)

  • Aim to create generative models that satisfy near-access freeness (NAF) criteria, which involves defining a safe function that maps a datapoint to a generative model trained without access to that datapoint, and measuring the divergence between the NAF model and the safe model using a suitable divergence measure. (Vyas, Kakade, and Barak 2023)

  • Strive to create a unified generalist framework capable of integrating the strengths of large language models (LLMs) with the specific requirements of vision-centric tasks, thereby enabling open-ended and customizable solutions for a wide range of vision-centric tasks. (Wenhai Wang et al. 2023)

  • Consider leveraging large language models to generate category-related descriptions along with structured graphs based on those descriptions, and subsequently implement Hierarchical Prompt Tuning (HPT) to enable simultaneous modeling of both structured and conventional linguistic knowledge for enhanced vision-language model performance. (Yubin Wang et al. 2023)

  • Consider employing the GPT-NER technique to bridge the gap between sequence labeling tasks like Named Entity Recognition (NER) and large language models (LLMs) by transforming the NER task into a text generation task that can be easily adapted by LLMs. Furthermore, they suggest implementing a self-verification strategy to mitigate the hallucination issue often encountered with LLMs. (Shuhe Wang et al. 2023)

  • Consider combining large language models (LLMs) with computer-aided diagnosis (CAD) networks for medical imaging to enhance the output of multiple CAD networks, such as diagnosis networks, lesion segmentation networks, and report generation networks, by summarizing and reorganizing the information presented in natural language text format. (Sheng Wang et al. 2023)

  • Carefully consider the role of semantic priors and input-label mappings in in-context learning, especially when working with large language models, as the ability to override semantic priors and learn input-label mappings emerges with model scale. (J. Wei et al. 2023)

  • Focus on developing a novel model called Graph-Grounded Pre-training and Prompting (G2P2) to address low-resource text classification problems, which involves jointly pre-training a graph-text model using three graph interaction-based contrastive strategies, followed by exploring handcrafted discrete prompts and continuous prompt tuning for downstream classification. (Z. Wen and Fang 2023)

  • Leverage the power of pre-trained image-text embeddings and fixed classname tokens to ensure robustness in your vision-language models, particularly when dealing with noisy labels. (C.-E. Wu et al. 2023)

  • Consider leveraging graph data to enhance the design of prompts in order to improve the effectiveness of the “pre-train, prompt, predict” training paradigm. (C. Wu et al. 2023)

  • Consider adopting the Prompt-Free Diffusion’ technique for text-to-image (T2I) research, which replaces traditional textual prompts with visual inputs, thereby reducing the need for time-consuming and subjective prompt engineering processes. (X. Xu et al. 2023)

  • Consider combining model compression methods with soft prompt learning strategies to optimize the accuracy-efficiency trade-off in large language models deployed on commodity hardware. (Zhaozhuo Xu et al. 2023)

  • Consider employing ChatGPT for diverse text summarization tasks, as it demonstrates strong performance comparable to traditional fine-tuning methods in terms of Rouge scores. (Xianjun Yang et al. 2023)

  • Consider using a universal continuous mapping framework like Uni-Fusion for handling diverse types of data in robotics, as it enables efficient encoding and generation of continuous surfaces, surface property fields, and other features without requiring extensive training. (Y. Yuan and Nuechter 2023)

  • Consider implementing AdaLoRA, a method that uses singular value decomposition to adaptively allocate the parameter budget among weight matrices according to your importance score, thereby improving the performance of parameter-efficient fine-tuning in large pre-trained language models. (Qingru Zhang et al. 2023)

  • Use GPT-4V as a generalist evaluator for vision-language tasks, as it shows promising agreement with humans across various tasks and evaluation methods, despite certain limitations. (Xinlu Zhang et al. 2023)

  • Consider implementing Ginsew, a novel method for protecting text generation models from being stolen through distillation, which involves injecting secret signals into the probability vector of the decoding steps for each target token, allowing for the detection of potential intellectual property infringements with minimal impact on the generation quality of protected APIs. (X. Zhao, Wang, and Li 2023)

  • Integrate Large Language Models (LLMs) into existing pre-trained vision-language (VL) models to enhance your ability to perform low-shot image classification tasks, particularly when dealing with limited or inaccessible training images. (Zhaoheng Zheng et al. 2023)

  • Carefully analyze the underlying factors causing object hallucination in large vision-language models, such as co-occurrence, uncertainty, and object position, before developing effective algorithms like LVLM Hallucination Revisor (LURE) to revise and improve the accuracy of generated descriptions. (Yiyang Zhou et al. 2023)

  • Focus on understanding and leveraging the neural collapse phenomenon in vision-language models to improve your generalization capabilities, particularly in class imbalance scenarios. (Z. Zhu et al. 2023)

  • Consider implementing a bi-level routing attention mechanism in your vision transformer models to achieve dynamic, query-aware sparsity, resulting in improved computational efficiency and performance. (Lei Zhu et al. 2023)

  • Carefully consider the choice of residual point sampling method for physics-informed neural networks (PINNs), as it greatly impacts the performance of PINNs in solving both forward and inverse problems of partial differential equations (PDEs). (C. Wu et al. 2023)

  • Carefully consider and compare various accuracy repair techniques when working with Binary Neural Networks (BNNs) to mitigate the significant accuracy loss caused by extreme quantization, ultimately leading to improved deployment on resource-constrained embedded systems. (Putter and Corporaal 2023)

  • Consider using a memory-augmented transformer architecture when dealing with language-guided video segmentation tasks, as it allows for efficient querying of the entire video with the language expression, while effectively capturing long-term context and avoiding visual-linguistic misalignment. (C. Liang et al. 2023)

  • Consider utilizing unsupervised representation learning (URL) techniques when working with point cloud data, as these methods can effectively handle various real-world tasks and significantly reduce the need for labeled data and manual annotations. (A. Xiao et al. 2023)

  • Carefully consider the role of memorization in your models, particularly when working with noisy datasets, and utilize appropriate techniques to mitigate its effects on model performance and generalization. (Rabin et al. 2023)

  • Carefully evaluate and optimize the quality of your pre-training data, model architecture, training approaches, and decoding strategies when developing large-scale pre-trained open-domain Chinese dialogue systems. (Yuxian Gu et al. 2023)

  • Utilise cross-task prototypes to model relationships between training tasks in episodic few-shot learning for event detection, enforcing prediction consistency among classifiers across tasks to enhance model robustness against outliers. (Xintong Zhang et al. 2023)

  • Consider incorporating optically reconfigurable supercomputers, specifically TPU v4, into your experimental designs to achieve significant improvements in scalability, availability, utilization, modularity, deployment, security, power efficiency, and overall performance when working with machine learning models. (Jouppi et al. 2023)

  • Differentiate between mechanical writing, which involves communicating existing information and can be performed effectively by machines, and sophisticated writing, which entails generating new insights through the writing process and requires critical thinking skills beyond the capabilities of current language generation models. (Bishop 2023)

  • Utilise the GradICON regulariser when conducting learning-based image registration. This technique involves penalising the Jacobian of the inverse consistency condition instead of the inverse consistency directly, leading to improved convergence, elimination of the requirement for careful scheduling of the inverse consistency penalty, production of spatially regular maps, and enhanced registration accuracy. (Rushmore et al. 2022)

  • Consider using multi-modal architectures that combine both visual and textual descriptors for extreme classification tasks involving millions of labels, as they can provide more accurate categorizations compared to traditional text-based or image-based methods. (A. Mittal et al. 2022)

  • Consider utilizing a multi-scale GAN-based model built on a tri-plane hybrid representation to effectively capture the geometric features of a single reference 3D shape across a range of spatial scales, allowing for the generation of diverse and high-quality 3D shapes potentially of different sizes and aspect ratios. (R. Wu and Zheng 2022)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (“Handbook of Digital Face Manipulation and Detection” 2022)

  • Consider using a novel CLIP-based spatio-textual representation for text-to-image generation tasks, allowing for greater control over the shapes of different regions/objects and your layout in a fine-grained manner. (Ackermann and Li 2022)

  • Focus on developing a scalable infrastructure that decouples model cost evaluation, search space design, and the NAS algorithm to effectively target various on-device ML tasks, while incorporating group convolution based inverted bottleneck (IBN) variants to optimize quality/performance trade-offs on ML accelerators. (Akin et al. 2022)

  • Focus on developing a joint embedding space for various modalities using image-paired data, rather than requiring all possible combinations of paired data, as this approach enables emergent capabilities and improves overall performance. (Alayrac et al. 2022)

  • Carefully examine the privacy implications of diffusion models, as they tend to memorize and reproduce individual training examples, potentially leading to privacy breaches and digital forgery issues. (H. Ali, Murad, and Shah 2022)

  • Optimize deep neural networks (DNNs) to inherently provide explanations that are both faithful summaries of the models and have clear interpretations for humans, rather than trying to optimize the explanation method itself. (Böhle, Fritz, and Schiele 2022)

  • Employ Prefix Conditioning to unify image-caption and image classification datasets for improved zero-shot recognition performance. (S. C. Y. Chan et al. 2022)

  • Consider incorporating a spatial self-attention layer within your transformer architecture to enhance 3D spatial understanding, allowing for improved language-conditioned spatial relation reasoning. (Shizhe Chen et al. 2022)

  • Utilize the three-pole signed distance function (3PSDF) for learning surfaces with arbitrary topologies, as it allows for easier field-to-mesh conversion using the classic Marching Cubes algorithm and outperforms previous state-of-the-art methods in various benchmarks. (Weikai Chen et al. 2022)

  • Use the Prompt-aligned Gradient (ProGrad) approach to effectively tune prompts in order to maintain alignment with general knowledge and prevent overfitting during few-shot learning. (Guangyi Chen et al. 2022)

  • Consider applying Multiple Instance Learning (MIL) techniques to aggregate and analyze multiple related images in conjunction with textual data, rather than relying solely on single image analysis. (H. W. Chung et al. 2022)

  • Utilise a novel positional encoding mechanism for physics-informed neural networks (PINNs) based on the eigenfunctions of the Laplace-Beltrami operator. This technique enables the creation of an input space for the neural network that accurately represents the geometry of a given object, allowing for improved solutions to forward and inverse problems involving partial differential equations. (Costabal, Pezzuto, and Perdikaris 2022)

  • Consider implementing a sparse version of causal attention mechanism in order to achieve low computational complexity when generating videos with increasing frames. (Couairon et al. 2022)

  • Consider using a Transformer-based model for Arbitrary Point cloud Upsampling (APU-SMOG) because it enables effective upsampling with any scaling factor, including non-integer values, with a single trained model. (Dell’Eva, Orsingher, and Bertozzi 2022)

  • Consider the impact of quantization error accumulation across time steps and the varying activation distributions across time steps when developing post-training quantization (PTQ) solutions for diffusion models. (Dettmers et al. 2022)

  • Consider developing efficient self-supervised learning (SSL) techniques for speech representation learning that balance generalizability and computation requirements, as measured by metrics like SUPERB score, MACs, and Params. (T. Feng et al. 2022)

  • Consider implementing GPTQ, a novel one-shot weight quantization technique based on approximate second-order information, to improve efficiency and accuracy in post-training quantization of large transformer models. (Frantar et al. 2022)

  • Consider utilizing ObjectFolder 2.0, a large-scale, multisensory dataset of common household objects in the form of implicit neural representations, to enhance the generalizability of your models to real-world scenarios. (R. Gao et al. 2022)

  • Employ a two-step approach consisting of visual-relation pre-training followed by prompt-based fine-tuning to effectively address the challenge of open-vocabulary scene graph generation (Ov-SGG) and enhance the models ability to predict visual relationships for unseen objects.’ (Tao He et al. 2022)

  • Employ counterfactual generation and contrastive learning in a joint optimization framework to enhance the generalizability of prompt learning for vision and language models. (Xuehai He et al. 2022)

  • Aim to develop models that enable the identification of physical parameters from just a single video, while maintaining interpretability and long-term prediction capabilities. (Hofherr et al. 2022)

  • Utilize neuro-symbolic approaches like VisProg to efficiently and effectively expand the scope of AI systems to serve the long tail of complex tasks that people may wish to perform. (Ziniu Hu et al. 2022)

  • Employ graph neural networks to analyze bitcoin address behavior, specifically by constructing a unified graph representation of address transactions, learning graph representations, and performing address classification. (Zhengjie Huang et al. 2022)

  • Utilise the Neyman (1923)s repeated sampling framework to statistically infer heterogeneous treatment effects discovered by generic machine learning algorithms in randomised experiments. (Imai and Li 2022)

  • Employ instance-aware prompt learning techniques to improve the accuracy and adaptability of pre-trained language models across diverse samples within a task. (F. Jin et al. 2022)

  • Consider implementing multi-modal prompt learning (MaPLe) when working with vision-language (V-L) models like CLIP, as it enables simultaneous adaptation of both language and vision branches, resulting in improved alignment between vision and language representations. (Khattak et al. 2022)

  • Consider implementing E-Branchformer, an enhanced version of Branchformer, which incorporates an effective merging method and additional point-wise modules to achieve state-of-the-art word error rates in automatic speech recognition tasks. (K. Kim et al. 2022)

  • Consider utilizing a novel method called “Primitive3D” for creating large-scale, diverse, and richly-annotated 3D object datasets through the assembly of randomly selected primitives. (Xinke Li et al. 2022)

  • Focus on improving the clustering of feature points and the adaptation to unseen tasks in few-shot medical segmentation, rather than simply increasing the number of prototypes. (Yiwen Li et al. 2022)

  • Consider incorporating causality-pruning knowledge prompts when working with pre-trained vision-language models to enhance your performance and adaptability across diverse domains. (Jiangmeng Li et al. 2022)

  • Consider developing a fully differentiable quantization method for vision transformers (ViT) that allows for the automatic learning of optimal bit-width allocations for different components within the transformer layers, taking into account the varying degrees of quantization robustness exhibited by those components. (Zhexin Li et al. 2022)

  • Consider implementing an end-to-end unsupervised speech recognition system like wav2vec-U 2.0, which eliminates the need for audio-side pre-processing and improves accuracy through better architecture, leading to improved unsupervised recognition results across multiple languages. (Haolin Liu et al. 2022)

  • Consider combining traditional digital signal processing (DSP) techniques with deep learning approaches to achieve improved noise-robustness and generalization in fundamental frequency (F0) estimation tasks. (Yisi Liu et al. 2022)

  • Use Subspace Prompt Tuning (Sub_PT) to mitigate overfitting issues in prompt tuning for vision-language models, while enhancing your generalization abilities through the incorporation of a Novel Feature Learner (NFL). (Chengcheng Ma et al. 2022)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Marsden, Döbler, and Yang 2022)

  • Consider employing variable-length subsampling techniques in conjunction with fixed-length subsampling strategies to effectively compress self-supervised speech models, thereby enhancing your efficiency and performance on downstream tasks. (Y. Meng et al. 2022)

  • Focus on developing differentiable approaches for re-basin, which enables the integration of any loss function and improves the efficiency and stability of the training process. (Peña et al. 2022)

  • Utilise a meta-learning based method called Meta-PDE’, which combines meta-learning and physics-informed neural networks (PINNs) to accelerate the solving of Partial Differential Equations (PDEs) without requiring a mesh or explicit supervision from ground truth data. (Tian Qin et al. 2022)

  • Consider developing a generalist agent like Gato, which utilizes a single neural network with the same set of weights to perform a wide range of tasks across different environments, thereby reducing the need for handcrafting policy models and increasing the amount and diversity of training data. (S. Reed et al. 2022)

  • Consider utilizing a novel compositional semantic mix (CoSMix) technique for unsupervised domain adaptation (UDA) in 3D LiDAR semantic segmentation tasks, as it effectively reduces domain shifts and outperforms existing state-of-the-art methods. (Saltori et al. 2022)

  • Ensure that your prompts are topically related to the task domain and calibrate the prior probability of label words to enhance the effectiveness of your language models. (Weijia Shi et al. 2022)

  • Consider leveraging pre-trained vision and language models such as CLIP and HuBERT to improve speech processing tasks, particularly when transcription costs are prohibitive. (Shih et al. 2022)

  • Consider using Direct Feedback Alignment (DFA) and specifically designed integer activation functions called pocket activations when developing algorithms for training Deep Neural Networks (DNNs) entirely with integer-only arithmetic, as this approach helps to overcome issues like overflow and improves compatibility across various platforms. (J. Song and Lin 2022)

  • Consider combining multiple strategies like point clustering, temporal consistency, translation equivariance, and self-supervision to develop robust unsupervised object detection models. (Yuqi Wang, Chen, and Zhang 2022)

  • Utilize a diffusion model for 3D novel view synthesis, specifically the 3DiM model, which uses a pose-conditional image-to-image diffusion model and a novel technique called stochastic conditioning to generate multiple views that are 3D consistent. (D. Watson et al. 2022)

  • Consider using compressed prompts in a Bayesian attribute framework to steer text generation towards desirable outcomes and away from undesirable ones, particularly in the context of toxicity reduction. (Wingate, Shoeybi, and Sorensen 2022)

  • Carefully analyze the impact of individual words and phrases within textual prompts on the generated images, as different linguistic categories (adjectives, nouns, etc.) consistently affect the image generation process differently. (Witteveen and Andrews 2022)

  • Consider using Wav2Seq, a novel self-supervised approach to pre-train both the encoder and decoder parts of encoder-decoder models for speech data, which involves generating a pseudo language as a compact discrete representation and formulating a self-supervised pseudo speech recognition task to transcribe audio inputs into pseudo subword sequences. (F. Wu et al. 2022)

  • Adopt a hierarchical optimal transport approach when comparing different neural network architectures, as it allows for simultaneous consideration of cell-level micro-architecture similarities and network-level macro-architecture differences. (Yeaton et al. 2022)

  • Consider utilizing range images rather than 3D point clouds for lidar data compression, as it allows for direct exploitation of lidar scanning patterns and improved compression efficiency. (X. Zhou et al. 2022)

  • Consider using prompt-learning based on knowledgeable expansion when working with short text classification tasks, as it allows for the integration of both the short text itself and external knowledge from open Knowledge Graphs like Probase to create more effective label words. (Yi Zhu et al. 2022)

  • Utilize Deep Gaussian Processes (DGPs) and scalable variational inference techniques to enhance the efficiency and effectiveness of Bayesian calibration of computer models, thereby enabling better handling of model complexity and reducing computational burdens. (Marmin and Filippone 2022)

  • Utilize self-supervised representation learning (SSRL) methods to effectively train deep neural networks (DNNs) without the need for extensive labeled datasets, thereby reducing the reliance on costly and time-consuming human annotation processes. (Ericsson et al. 2022)

  • Consider using skip connections in your encoder-decoder models when working with unorganized sets of 3D feature maps, as this helps to preserve fine geometric details from the given partial input cloud and leads to improved completion accuracy and reduced memory occupancy. (Yida Wang et al. 2022)

  • Carefully consider the potential impact of scale disparities between objective functions when combining them in a composite objective function for physics-informed neural networks, as improper scaling can lead to difficulties in learning and convergence. (Basir and Senocak 2022)

  • Utilise the Stochastic Physics-Informed Neural Ordinary Differential Equations (SPINODE) framework to effectively learn the hidden physics within Stochastic Differential Equations (SDEs) by combining the principles of neural ordinary differential equations (Neural ODEs) and physics-informed neural networks (PINN) to approximate the weights and biases within the neural network representing g(x) from state trajectory data. (O’Leary, Paulson, and Mesbah 2022)

  • Choose test functions of the lowest polynomial degree and use quadrature formulas of suitably high precision to achieve a high decay rate of the error in Variational Physics Informed Neural Networks (VPINN) for smooth solutions. (Berrone, Canuto, and Pintore 2022)

  • Consider utilising Meta-Weight-Net, a novel method that enables the adaptive learning of an explicit weighting function directly from data, thereby improving the robustness of deep neural networks trained on biased data. (K. Kawaguchi, Bengio, and Kaelbling 2022)

  • Consider using the AdaIN-based method and a design of decoders to decouple geometry and appearance embedded in the tri-plane, enabling intuitive geometry editing by semantic masks. (S.-Y. Chen et al. 2022)

  • Consider implementing the Dendritic Gated Network (DGN) model, which combines dendritic “gating” with local learning rules to offer a biologically plausible alternative to backpropagation, resulting in improved efficiency, reduced forgetting, and superior performance across various tasks compared to traditional artificial networks. (Sezener et al. 2021)

  • Consider using the Automatic Relevance Determination (ARD) model for non-linear regression tasks, as it allows for the introduction of multiple regularisation constants, one associated with each input, which helps to identify and eliminate irrelevant variables, thereby improving model performance. (Smith and Gasper 2021)

  • Consider deploying tools initially developed for low-latency applications in science for low-power applications, focusing on ML for FPGAs and ASICs as energy efficient hardware architectures. (Tran et al. 2021)

  • Consider using symmetry regularization (SymReg) and saturating nonlinearity (SatNL) techniques to enhance the robustness of neural networks against quantization, leading to improved performance across various bit-widths and quantization schemes. (J.-W. Jang et al. 2021)

  • Consider leveraging large datasets in resource-rich languages to improve the efficiency and accuracy of your models for resource-poor languages, particularly through effective pre-training and fine-tuning techniques. (Orihashi et al. 2021)

  • Consider utilizing neural implicit representations instead of explicit geometric ones for object-object interaction problems, as it may lead to a paradigm shift and open doors to radically different approaches. (Andrews and Erleben 2021)

  • Utilise a self-adaptive loss balanced method for physics-informed neural networks (lbPINNs) to enhance your approximation capabilities. (L.-S. Zhang et al. 2021)

  • Carefully fine-tune large text-to-image diffusion models using a few images of a subject and a unique identifier, along with an autogenous class-specific prior preservation loss, to effectively generate novel photorealistic images of the subject in diverse scenes, poses, views, and lighting conditions while preserving its key features. (Abdal et al. 2021)

  • Consider leveraging cross-modal information to improve the efficiency and efficacy of few-shot learning systems, particularly in cases where traditional unimodal approaches may struggle to accurately characterize complex concepts. (Afham et al. 2021)

  • Consider implementing a Gradient Switching Strategy (GSS) when dealing with noisy labels in deep learning models. This strategy involves creating a gradient direction pool for each sample, which contains all-class gradient directions with varying probabilities. During training, the gradient direction pool is updated iteratively, assigning higher probabilities to potential principal directions for high-confidence samples while forcing uncertain samples to explore in different directions instead of misleading the model in a fixed direction. This approach helps mitigate (Bar, Koren, and Giryes 2021)

  • Adopt a hardware-aware Neural Architecture Search (HW-NAS) approach when developing deep learning models for resource-constrained platforms, as it enables the creation of efficient architectures that balance accuracy and hardware constraints. (Benmeziane et al. 2021)

  • Conduct multiple runs of your deep learning experiments using various random seeds to assess the impact of randomness on performance outcomes, as this can significantly affect the perceived significance of results. (M. Caron et al. 2021)

  • Consider employing a combination of context-aware spatial-semantic alignment and mutual 3D-language masked modeling when developing 3D-language pre-training techniques for improved cross-modal information exchange and reduced relational ambiguities. (D. Z. Chen et al. 2021)

  • Utilize the OpenPrompt framework when studying prompt-learning, as it offers a unified, easy-to-use, and extensible platform that simplifies the process of combining different pre-trained language models, task formats, and prompting modules. (N. Ding et al. 2021)

  • Consider implementing a “background interpretation scheme” and a “context grading scheme with tailored positive proposals” when developing a detection prompt (DetPro) system for open-vocabulary object detection based on a pre-trained vision-language model. (Han Fang et al. 2021)

  • Consider using a graphics-inspired factorization technique when working with Neural Radiance Fields (NeRF) systems, as it enables efficient caching and reduces memory complexity, ultimately allowing for high-quality photorealistic rendering at 200 frames per second on consumer-grade hardware. (Garbin et al. 2021)

  • Develop a novel trustworthy multimodal classification algorithm called “Multimodal Dynamics” that dynamically evaluates both the feature-level and modality-level informativeness for different samples, allowing for trustworthy integration of multiple modalities. (Gawlikowski et al. 2021)

  • Consider adopting a variational Bayesian approach for unsupervised similarity learning in atlas-based non-rigid medical image registration, as it enables the estimation of a data-specific similarity metric with relatively little data, improves robustness through the approximate variational posterior of the transformation parameters, and allows for the quantification of uncertainty associated with the output. (Grzech et al. 2021)

  • Consider using EigenGAN, a novel approach that enables unsupervised mining of interpretable and controllable dimensions from different generator layers within a Generative Adversarial Network (GAN), allowing for manipulation of specific semantic attributes in synthesized images. (Zhenliang He, Kan, and Shan 2021)

  • Consider combining multiple sources of data, such as WiFi signals, inertial measurements, and floor plans, to achieve higher levels of accuracy and density in estimating location histories in indoor environments. (Herath et al. 2021)

  • Consider utilizing the Convolutional Point Transformer (CpT) architecture for effectively handling unstructured 3D point cloud data, as it demonstrates superior performance compared to existing attention-based Convolutional Neural Networks and previous 3D point cloud processing transformers. (Kaul et al. 2021)

  • Consider using tapered fixed-point numerical format for your TinyML models, as it provides better dynamic range and precision adjustment capabilities compared to traditional fixed-point formats, resulting in higher inference accuracy and lower quantization errors. (Langroudi et al. 2021)

  • Consider utilizing residual energy-based models (R-EBMs) alongside traditional auto-regressive models for end-to-end speech recognition tasks, as it helps bridge the gap between the model and data distributions, leading to significant improvements in word error rate reductions and utterance-level confidence estimation performances. (Qiujia Li et al. 2021)

  • Consider integrating causal reasoning into data-free quantization processes to enhance the accuracy and efficiency of model compression techniques. (Yuang Liu et al. 2021)

  • Aim for pareto-optimality in your deep learning models, balancing model quality against factors such as model size, latency, resource requirements, and environmental impact. (Menghani 2021)

  • Carefully consider the trade-off between computational efficiency and memory constraints when implementing out-of-core neural networks on microcontroller units (MCUs), taking advantage of parallelism opportunities and optimizing tile sizes to minimize swapping overhead. (Hongyu Miao and Lin 2021)

  • Consider implementing multi-task learning for end-to-end automatic speech recognition (ASR) systems, specifically focusing on jointly learning word confidence, word deletion, and utterance confidence, as this approach leads to improvements in confidence metrics (such as NCE, AUC, and RMSE) without requiring an increase in the model size of the confidence estimation module. (D. Qiu et al. 2021)

  • Consider using Latent Optimization of Hairstyles via Orthogonalization (LOHO) for hairstyle transfer, as it enables users to synthesize novel photorealistic images by manipulating hair attributes either individually or jointly, achieving superior performance compared to existing approaches. (Saha et al. 2021)

  • Consider integrating CLIP (a Contrastive Language-Image Pre-training model) as the visual encoder within various Vision-and-Language (V&L) models, as it demonstrates significant improvements in performance when compared to traditional visual encoders trained on smaller sets of manually-annotated data. (S. Shen et al. 2021)

  • Carefully evaluate the performance of deep learning models for tabular data alongside established methods like XGBoost, considering factors such as accuracy, efficiency, and hyperparameter tuning, before deciding on the optimal approach for your particular application. (Shwartz-Ziv and Armon 2021)

  • Consider integrating positional information into the learning process of transformer-based language models using the novel Rotary Position Embedding (RoPE) method, which encodes absolute position with a rotation matrix and explicitly incorporates relative position dependencies within the self-attention formulation. (Jianlin Su et al. 2021)

  • Consider utilizing the proposed “Knowledge Evolution” (KE) approach when working with deep learning models on relatively small datasets. This involves splitting the model into two hypotheses—a fit-hypothesis’ and a ‘reset-hypothesis’. The ‘fit-hypothesis’ is evolved by perturbing the ‘reset-hypothesis’ over several generations, leading to improved performance and reduced inference costs.’ (Taha, Shrivastava, and Davis 2021)

  • Combine Physics Informed Neural Networks (PINNs) with traditional analytical methods like Airy stress functions and Fourier series to achieve highly accurate and efficient solutions for difficult biharmonic problems of elasticity and elastic plate theory. (Vahab et al. 2021)

  • Carefully consider the potential impact of dataset bias on model-based candidate generation systems and explore methods such as random negative sampling and fine-tuning to mitigate these biases. (Virani et al. 2021)

  • Carefully choose the “batch” in BatchNorm to optimize model performance, taking into account various factors such as normalization statistics, batch size, and potential domain shifts. (Yuxin Wu and Johnson 2021)

  • Employ the Semantic Point Generation (SPG) technique when dealing with unsupervised domain adaptation (UDA) for LiDAR-based 3D object detection, particularly when faced with issues arising from deteriorating point cloud quality due to varying environmental conditions like weather. (Q. Xu et al. 2021)

  • Use a self-training pipeline called ST3D for unsupervised domain adaptation on 3D object detection tasks, which involves pre-training the 3D detector on the source domain with a random object scaling strategy, followed by iterative improvement on the target domain through pseudo label updating with a quality-aware triplet memory bank and model training with curriculum data augmentation. (Jihan Yang et al. 2021)

  • Utilize semi-automatic annotation techniques to condense large volumes of audio data, allowing for more efficient and accurate identification of distinct species vocalizations. (Zwerts et al. 2021)

  • Consider using the Xtensa LX6 microprocessor within the ESP32 SoC for neural network applications, particularly in situations requiring low power consumption and fast processing speeds. (WANG et al. 2021)

  • Consider using a data-driven approach to learn a deformable model for 3D garments from monocular images, rather than relying solely on physics-based simulations, in order to avoid high computational costs and the simulation-to-real gap. (S. Bang, Korosteleva, and Lee 2021)

  • Focus on developing a deep Linear Discriminant Analysis (LDA)-based neuron/filter pruning framework that is aware of both class separation and holistic cross-layer dependency, allowing for efficient and effective pruning of unnecessary features in deep neural networks. (Q. Tian, Arbel, and Clark 2021)

  • Develop a deep neural network called “Point Transformer” that operates directly on unordered and unstructured point sets, using a local-global attention mechanism to capture spatial point relations and shape information, and integrating a SortNet module to ensure input permutation invariance. (N. Engel, Belagiannis, and Dietmayer 2021)

  • Adopt a comprehensive training methodology for TinyML models, taking into account analog non-idealities such as conductance drift, read/write noise, and fixed analog-to-digital converter gains, to minimize accuracy loss when deploying them on analog compute-in-memory systems. (Dazzi et al. 2021)

  • Carefully balance the benefits of domain decomposition in reducing the complexity of learned solutions against the potential drawbacks of having less training data per subdomain, which could result in overfitting and reduced generalizability. (S. Cai et al. 2021)

  • Consider implementing an end-to-end edge device application (TinyML based) for real-time predictive maintenance (Fault Detection and Remaining Useful Life) of Solenoid Valves (SV), using a custom-built intelligent electronic product that encapsulates data acquisition, feature extraction, and inference in a tiny embedded package. (Amrane et al. 2021)

  • Consider developing a modular framework for predicting video memorability, which involves processing input videos in a tiered manner, with each module focusing on a specific aspect of the visual content, such as raw encoding, scene understanding, event understanding, and memory consolidation. (“Augmented Cognition” 2021)

  • Pay close attention to the benchmarking process, avoid direct hyperparameter optimization on the test set, and use a shared train/validation/test split for proper evaluation settings when comparing state-of-the-art methods in entity alignment tasks. (Berrendorf, Wacker, and Faerman 2021)

  • Utilize the Stanford Sentiment Treebank and the Recursive Neural Tensor Network (RNTN) model to achieve superior results in sentiment analysis tasks, particularly in capturing the nuances of negation and its scope across various tree levels for both positive and negative phrases. (Beam et al. 2021)

  • Consider using the CNewSum dataset when developing and testing Chinese news summarization models, as it offers a large-scale collection of human-written summaries along with adequacy and deducibility scores to guide the development of more human-friendly summaries. (Danqing Wang et al. 2021)

  • Consider using a two-stage approach when attempting to automatically generate 3D human motions from text, combining text2length sampling and text2motion generation, and utilizing motion snippet code as an internal motion representation to improve the accuracy and diversity of the resulting motions. (S. Ghorbani et al. 2021)

  • Utilize a multi-objective constrained neural architecture search (NAS) algorithm, specifically μNAS, to optimize for multiple objectives simultaneously in the context of microcontroller-level architectures. (Liberis, Dudziak, and Lane 2021)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Bender et al. 2021)

  • Consider using the hyperbolic space for interpolative data augmentation, as it captures the complex geometry of input and hidden state hierarchies better than its contemporaries, leading to consistent outperformance of state-of-the-art data augmentation techniques across multiple domains. (Sawhney et al. 2021)

  • Utilize a deep multimodal multilabel learning (DMML) approach to detect the existence of multiple illicit drugs from suspect illicit drug trafficking events (IDTEs) on Instagram, incorporating both text and image data for improved accuracy. (C. Hu et al. 2021)

  • Consider incorporating product seasonal relevance into search ranking algorithms to improve search results and enhance customer satisfaction. (Haode Yang et al. 2021)

  • Develop an iterative learning paradigm consisting of a label aggregation stage and a label correction stage to improve the accuracy of fraud detection models trained on multi-sourced noisy annotations. (Chuang Zhang et al. 2021)

  • Focus on developing large, multilingual, and high-quality datasets for multimodal learning, as exemplified by the presented Wikipedia-based Image Text (WIT) Dataset, which offers superior performance compared to smaller, monolingual datasets. (Srinivasan et al. 2021)

  • Consider developing a lifelong user representation learning system, named Conure, which allows for continual learning of user profiles across multiple tasks without forgetting previous information. (F. Yuan et al. 2021)

  • Utilize the AutoCTS algorithm to automatically identify highly competitive spatiotemporal (ST) blocks and forecasting models with heterogeneous ST-blocks connected using diverse topologies, thereby improving the efficiency and accuracy of correlated time series forecasting. (Xinle Wu et al. 2021)

  • Focus on developing models that effectively capture both explicit and implicit feature interactions, while remaining computationally efficient and scalable for practical implementation. (Ruoxi Wang et al. 2021)

  • Utilise Physics-Informed Neural Networks (PINNs) for solving Partial Differential Equations (PDEs) as they offer advantages like being mesh-free, breaking the curse of dimensionality, and providing a direct strong form approach that avoids truncation errors and numerical quadrature errors of variational forms. (Lu Lu et al. 2021)

  • Focus on developing surrogate models for complex systems that are robust to model misspecification and capable of handling nonlinear phenomena through appropriate approximations. (Bhattacharya et al. 2021)

  • Focus on developing and utilizing benchmarks that accurately assess the reasoning abilities of Visual Question Answering (VQA) models, rather than solely relying on overall in-domain accuracy measurements, which may be influenced by dataset biases. (Sverrisson et al. 2020)

  • Build large-scale, diverse, and representative datasets for training deep learning models to improve the accuracy of no-reference video quality assessment (NR-VQA) predictions. (Sverrisson et al. 2020)

  • Utilize a combination of pre-existing text-to-image models and unsupervised learning techniques on unlabelled video data to create text-to-video models, thereby avoiding the need for paired text-video data and improving overall model performance. (Girish, Singh, and Ralescu 2020)

  • Utilize a fine-tuned deep residual network (ResNet) for time series classification tasks, particularly when dealing with small amounts of labeled data. (Rakhshani et al. 2020)

  • Develop a deep learning framework specifically tailored for motion retargeting between skeletons with different structures, leveraging the concept of a “primal skeleton” and introducing novel differentiable convolution, pooling, and unpooling operators that are aware of the skeletons hierarchical structure and joint adjacency.’ (Aberman et al. 2020)

  • Use Shapley value to evaluate the contribution of operations in neural architecture search, rather than relying solely on the magnitude of architecture parameters updated by gradient descent. (Ancona, Öztireli, and Gross 2020)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Bar-On et al. 2020)

  • Consider using the Adam optimization algorithm instead of Stochastic Gradient Descent (SGD) for Binary Neural Networks (BNNs) due to its superior handling of the rugged loss surface and its ability to revitalize dead’ weights caused by activation saturation, leading to improved generalization ability.’ (Bethge et al. 2020)

  • Consider incorporating visual information into text classification tasks by leveraging vision-language pre-training models (VL-PTMs) through a novel method called Visual Prompt Tuning (VPT), which generates visual prompts for category names and adds them to the alignment process, leading to improved performance in both zero-shot and few-shot settings. (T. B. Brown et al. 2020)

  • Consider implementing a performance-aware mutual knowledge distillation (PAMKD) approach for neural architecture search, where knowledge generated by model A is allowed to train model B only if the performance of A is better than B. (Ting Chen et al. 2020)

  • Utilise a unified perspective to analyse the expressive power and inductive bias of Implicit Neural Representations (INRs), leveraging results from harmonic analysis and deep learning theory. (D’Amour et al. 2020)

  • Carefully distinguish between expressivity and learnability when attempting to apply neural networks to causal inference problems, recognizing that even highly expressive neural networks may struggle to accurately capture the underlying causal relationships due to limitations in learnability. (Falcon and Cho 2020)

  • Carefully examine the relationship between the models choice of prices and what guests actually prefer, and ensure that the model takes into account the “cheaper is better” principle when ranking listings. (Haldar et al. 2020)

  • Develop specialized verification methods for quantized neural networks, taking into account the more complex semantics caused by quantization, rather than relying solely on methods designed for standard networks. (Henzinger, Lechner, and Žikelić 2020)

  • Consider using a pose-conditioned StyleGAN2 latent space interpolation technique for generating highly realistic and accurate try-on images, which involves optimizing for interpolation coefficients per layer to ensure a smooth combination of body shape, hair, skin color, and garment details. (Jialu Huang, Liao, and Kwong 2020)

  • Consider incorporating an inference-time label-preserving target projections technique to enhance the generalizability of machine learning models trained on a set of source domains to unseen target domains with different statistics. (Zeyi Huang et al. 2020)

  • Consider integrating the LP-MDN into the LPCNet vocoder to achieve higher quality synthetic speech by enabling the autoregressive neural vocoder to structurally represent the interactions between the vocal tract and vocal source components. (M.-J. Hwang et al. 2020)

  • Consider implementing differentiable neural architecture transformation techniques to overcome the limitations of existing Neural Architecture Transformers (NATs). (D.-G. Kim and Lee 2020)

  • Consider utilizing a low-rank representation of Kronecker factored eigendecomposition to reduce the space complexity of MND from O(N^3) to O(L^3), where L is the chosen low-rank dimension instead of parameter space lying in high dimensional N manifolds. (Jongseok Lee et al. 2020)

  • Consider implementing multipoint quantization for post-training quantization, which approximates a full-precision weight vector using a linear combination of multiple vectors of low-bit numbers, allowing for greater precision levels for important weights and avoiding the need for specialized hardware accelerators required by traditional mixed precision methods. (Xingchao Liu et al. 2020)

  • Incorporate scene text as a third modality in cross-modal retrieval tasks to enhance the accuracy and efficiency of the retrieval process. (Mafla et al. 2020)

  • Utilise the concept of redundancy among parameter groups within neural networks, leveraging rate-distortion theory to identify permutations that lead to functionally equivalent, yet easier-to-quantize networks. (Martinez et al. 2020)

  • Consider integrating machine learning techniques with existing scientific models to create a more robust and efficient framework for understanding complex phenomena. (Rackauckas et al. 2020)

  • Track MLPerf Mobiles benchmark tasks, accuracy metrics, quality thresholds, rules, etc., to present industry-relevant evaluations that practitioners can adopt to bridge the gap between research and practice.’ (Reddi et al. 2020)

  • Consider using a hybrid neural network architecture (HyNNA) for NVS-based surveillance applications, which combines dual-polarity event channels and CNN architectures for classification, resulting in significant improvements in accuracy and efficiency. (Singla et al. 2020)

  • Develop a highly efficient learning-based method for computing good approximations of optimal sparse codes in a fixed amount of time, assuming that the basis vectors of a sparse coder have been trained and are being kept fixed. (Yuhai Song et al. 2020)

  • Focus on developing a scalable, automated, and flexible data classification system that combines multiple data signals, machine learning, and traditional fingerprinting techniques to effectively manage and protect sensitive data within large organizations. (Tanaka, Sapra, and Laptev 2020)

  • Consider utilizing Sparse Point-Voxel Convolution (SPVConv) for efficient 3D architectures, which combines the benefits of point-based and voxel-based methods, preserving fine details even in large outdoor scenes. (H. Tang et al. 2020)

  • Carefully evaluate the impact of varying the inputs to the transformer on the exact match scores for different query types, particularly considering the trade-off between scalability and accuracy. (Thorne et al. 2020)

  • Integrate an external fact memory into a neural language model, allowing for improved performance on knowledge-intensive question-answering tasks and the ability to update and manipulate symbolic representations without retraining the entire model. (Verga et al. 2020)

  • Consider implementing a deep interest with hierarchical attention network (DHAN) for click-through rate prediction tasks, as it demonstrates improved accuracy over existing methods due to its ability to effectively model user interests across multiple dimensions and hierarchical levels. (Weinan Xu et al. 2020)

  • Consider implementing Mixed Negative Sampling (MNS) in your two-tower neural network models for large corpus item retrieval in recommendations, as it effectively addresses the issue of selection bias inherent in traditional batch negative sampling methods. (Ji Yang et al. 2020)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (J. Nelson 2020)

  • Carefully consider the potential impact of non-determinism in your machine learning models, particularly in safety-critical applications, as even minor sources of randomness can lead to significant changes in model performance on specific subsets of the data. (Alahmari et al. 2020)

  • Consider using Quantization Guided Training (QGT) as a regularizer-based approach for quantization-aware training (QAT) in deep neural networks, as it offers advantages such as improved stability, ease of implementation, and compatibility with various training pipelines. (Y. Choi, El-Khamy, and Lee 2020)

  • Consider implementing adaptive sparse backpropagation algorithms, such as TinyProp, when working with deep neural networks on resource-limited devices, as it offers improved efficiency and comparable accuracy to traditional backpropagation methods. (Xu Sun et al. 2020)

  • Consider using time-varying speaker representation for one-shot voice conversion, as opposed to fixed-size speaker representation, to better capture the dynamic nature of speech signals and reduce information loss. (Ishihara and Saito 2020)

  • Consider utilizing the proposed “Language Model Based Data Augmentation” (LAMBADA) technique when dealing with limited labeled data in text classification tasks. This method leverages a pre-trained language model to create new labeled data, which is then filtered using a classifier trained on the original data. By doing so, researchers can potentially enhance your classifiers performance, surpass current state-of-the-art data augmentation methods, and provide an attract (Marivate and Sefara 2020)

  • Consider using a time-variant deep feed-forward neural network architecture like ForecastNet for multi-step-ahead time-series forecasting, as it allows for better modeling of dynamics at a range of scales and resolutions compared to traditional time-invariant architectures. (Dabrowski, Zhang, and Rahman 2020)

  • Aim to minimise the extent to which prior assumptions about physical systems impose structure on the machine learning system, allowing for greater flexibility and potential for discovery. (Iten et al. 2020)

  • Focus on developing dynamic graph representation learning algorithms that effectively combine structural and temporal self-attention mechanisms to accurately capture the complexities of evolving graph structures. (Sankar et al. 2020)

  • Consider and account for both algorithmic and implementation-level non-deterministic factors (NI-factors) when evaluating deep learning (DL) systems, as these factors can significantly impact model performance and training time. (Pham et al. 2020)

  • Carefully select appropriate machine learning algorithms to optimize brain tumor segmentation, progression assessment, and overall survival prediction in the context of the BRATS challenge. (Zwanenburg et al. 2020)

  • Utilise the Least Absolute Deviation based PINN (LAD-PINN) and the two-stage Median Absolute Deviation based PINN (MAD-PINN) to accurately reconstruct solutions and recover unknown parameters in Partial Differential Equations (PDEs) even when faced with highly corrupted data. (Maziar Raissi, Yazdani, and Karniadakis 2020)

  • Focus on exploring a broad range of candidate operations, rather than limiting themselves to a predefined subset, and utilize efficient search strategies like progressive pruning and replacement to navigate the large search space effectively. (Laube and Zell 2019)

  • Consider applying quantization-aware training during the fine-tuning phase of BERT to effectively compress the model by 4x with minimal accuracy loss, potentially improving efficiency in production environments. (Zafrir et al. 2019)

  • Consider utilizing Socratic Models (SMs) as a modular framework to combine multiple pretrained models through language-based exchanges, allowing them to perform new downstream multimodal tasks without requiring additional training or fine-tuning. (Abuzaid et al. 2019)

  • Investigate the possibility of achieving energy savings in the computational path of deep neural network (DNN) hardware accelerators through the introduction of approximate arithmetic operators, without requiring time-consuming retraining processes. (Mrazek et al. 2019)

  • Optimize for measured quantities such as inference time, rather than focusing solely on theoretical computational efficiency metrics, when developing efficient network designs for deep learning computer vision applications. (Cubuk et al. 2019)

  • Consider implementing a k-quantile quantization method with balanced (equal probability mass) bins for neural networks, as it is particularly suitable for handling bell-shaped distributions commonly found in these systems. (Baskin et al. 2019)

  • Consider utilizing Ensemble Knowledge Distillation (EKD) for enhancing the classification performance and model generalization of compact networks. By distilling knowledge from multiple teacher networks into a compact student network via an ensemble architecture, EKD allows for increased heterogeneity in feature learning and improved prediction quality. (Asif, Tang, and Harrer 2019)

  • Consider using soft pseudo-labels rather than hard ones in order to allow students to distill richer information from teachers, prevent over-fitting to potentially incorrect predictions, and maintain flexibility in dealing with ambiguous cases. (Berthelot et al. 2019)

  • Consider implementing a teacher-student learning paradigm in your studies, where the teacher network generates pseudo-labels to optimize the student network. This approach enables models to leverage massive amounts of unlabeled data based on a smaller portion of labeled data, potentially reducing the need for costly and time-consuming manual annotation processes. (Berthelot et al. 2019)

  • Focus on maintaining rich information flow within the network rather than relying on complex approximation methods and training tricks when developing Binary Neural Networks (BNNs). (Bethge et al. 2019)

  • Focus on developing efficient 8-bit quantization techniques for Transformer neural machine language translation models, specifically by leveraging high-performance libraries like Intel® Math Kernel Library (MKL) matrix multiplication kernels optimized with INT8/VNNI instructions, to improve inference efficiency while maintaining minimal drops in BLEU score accuracy. (Bhandare et al. 2019)

  • Consider leveraging large amounts of unlabelled data in the wild to address the data-free knowledge distillation problem, rather than attempting to generate images solely from the teacher network. (Bhardwaj, Suda, and Marculescu 2019)

  • Focus on developing practical approaches for unlearning in machine learning systems, specifically through the use of data sharding and slicing techniques, in order to balance the need for accurate models with the growing demand for data privacy and protection. (Bourtoule et al. 2019)

  • Utilise adaptive estimation techniques when measuring mutual information in deep neural networks, as they allow for more accurate evaluation of different activation functions and reveal varying degrees of compression depending on the specific activation function employed. (Chelombiev, Houghton, and O’Donnell 2019)

  • Focus on developing a structured Bayesian compression architecture for deep neural networks, incorporating a mixture of sparsity inducing priors and structured sparsity learning techniques, to enable efficient and accurate model compression for mobile-enabled devices in connected healthcare. (Sijia Chen et al. 2019)

  • Focus on minimizing the distribution gap between the weights inherited from the supernet and the weights trained with stand-alone networks in order to achieve more accurate evaluations and improved overall performance in neural architecture search. (Yukang Chen et al. 2019)

  • Extend and adapt transductive zero-shot learning and generalized zero-shot learning to 3D point cloud classification, develop a novel triplet loss that takes advantage of unlabeled test data, and conduct extensive experiments to establish state-of-the-art results on multiple 3D datasets. (Cheraghian et al. 2019)

  • Utilize a cost-aware channel sparse selection (C2S2) methodology when attempting to simplify deep neural networks. This method involves adding a pruning layer to a pre-trained model, allowing for a two-phase optimization process that operates with an end-to-end differentiable network. By progressively performing the pruning task layer-wise and adhering to a sparsity criterion, the C2S2 method favors pruning more channels while developing (C.-Y. Chiu, Chen, and Liu 2019)

  • Carefully examine the potential impact of the “Co-adaptation Problem” and “Matthew Effect” on your neural architecture search (NAS) models, and consider implementing techniques such as “grouped operation dropout” to address these issues and improve model performance. (Chu et al. 2019)

  • Consider using an Image-specific Prompt Learning (IPL) method when working with generative model adaptation, as it allows for more precise and diversified adaptation directions, ultimately resulting in higher quality and more varied synthesized images. (Clouâtre and Demers 2019)

  • Consider implementing a novel inheritance and exploration knowledge distillation framework (IE-KD) to effectively train a student network by partially following the knowledge from the teacher network while also exploring for new knowledge that complements the teacher network. (Chunfeng Cui et al. 2019)

  • Consider using Global Sparse Momentum Stochastic Gradient Descent (GSM-SGD) for pruning very deep neural networks, as it offers benefits including automatic discovery of appropriate per-layer sparsity ratios, end-to-end training, no need for time-consuming re-training processes post-pruning, and enhanced ability to identify “winning tickets” that have benefited from favorable initial conditions. (X. Ding et al. 2019)

  • Consider implementing a resource-aware, efficient weight quantization framework like REQ-YOLO for object detection tasks on FPGAs, which combines software and hardware-level optimization opportunities and enables real-time, highly-efficient implementations. (C. Ding et al. 2019)

  • Consider using BigBiGAN, a modified version of the BigGAN model, for unsupervised representation learning, as it outperforms previous approaches in generating high-quality images and accurately representing semantic features. (J. Donahue and Simonyan 2019)

  • Consider using spatial relation modeling when working on vision-and-language reasoning tasks, as it helps to maintain more spatial context and focus attention on essential visual regions for reasoning. (L. Dong et al. 2019)

  • Utilise LayerDrop’, a form of structured dropout, to effectively manage overparameterised transformer networks. This method enables efficient pruning at inference time, allowing for the selection of sub-networks of any depth from one large network without requiring fine tuning, thereby reducing computational demands and mitigating overfitting risks.’ (A. Fan, Grave, and Joulin 2019)

  • Carefully examine the mean activation shift (MAS) in your neural networks, particularly in layers with fewer parameters, as it can significantly contribute to quantization errors and lead to decreased network performance. (Finkelstein, Almog, and Grobman 2019)

  • Carefully consider the potential effects of pruning on interpretability when applying pruning techniques to neural networks, as pruning may affect the interpretability of the model depending on the specific pruning method used. (Frankle and Bau 2019)

  • Consider utilizing the UV-Net neural network architecture and representation for operating directly on Boundary representation (B-rep) data from 3D CAD models, as it effectively addresses the challenges posed by the complexity of the data structure and its support for both continuous non-Euclidean geometric entities and discrete topological entities. (Jun Gao et al. 2019)

  • Carefully consider the choice of feature distribution when studying high-dimensional ridgeless least squares interpolation, as it can lead to the recovery of several phenomena observed in large-scale neural networks and kernel machines, including the “double descent” behavior of the prediction risk and the potential benefits of overparametrization. (Hastie et al. 2019)

  • Consider the use of Pruning-Aware Merging (PAM) for efficient multitask inference, as it enables “merge & prune” for reducing computation costs across different subsets of tasks. (Xiaoxi He et al. 2019)

  • Redefine latent weights as inertia and adopt the Binary Optimizer (Bop) for better understanding and optimization of Binarized Neural Networks (BNNs). (Helwegen et al. 2019)

  • Utilize the proposed ImageNet-C and ImageNet-P datasets to comprehensively assess the robustness of neural networks against common corruptions and perturbations, thereby enhancing overall network resilience and generalizability. (Hendrycks and Dietterich 2019)

  • Consider creating adversarially filtered datasets to expose and measure the vulnerabilities of machine learning models, particularly in cases where there might be spurious cues leading to inaccurate performance estimates. (Hendrycks et al. 2019)

  • Consider implementing natural compression (C_nat) as a novel, efficient, and theoretically sound compression technique for distributed deep learning tasks, which can lead to significant reductions in communication costs without compromising the accuracy of the model. (Horvath et al. 2019)

  • Consider multiple types of loss functions simultaneously during channel pruning of deep neural networks, specifically focusing on reconstruction error, classification loss, and feature and semantic correlation loss, to optimize model performance while reducing model complexity. (Yiming Hu et al. 2019)

  • Consider incorporating a low-rank constraint when working with multivariate data, as it can lead to significant improvements in efficiency and interpretability. (Humbert et al. 2019)

  • Consider implementing Network Implosion, a technique that involves static layer pruning and retraining of residual networks, to effectively compress models without compromising accuracy. (Ida and Fujiwara 2019)

  • Utilise a large amount of online handwriting data to train your line recogniser in an offline handwritten text recognition (HTR) system, rather than rely solely on manual labelling of handwritten text lines in images. (Ingle et al. 2019)

  • Utilise a two-stage learning framework for TinyBERT, which involves a general distillation phase followed by a task-specific distillation phase. This approach allows TinyBERT to capture both general-domain and task-specific knowledge from BERT, thereby enabling it to achieve high levels of performance while remaining computationally efficient. (X. Jiao et al. 2019)

  • Consider utilizing self-supervised learning methods for visual feature extraction from large-scale unlabelled datasets, as it allows for effective feature learning without requiring extensive manual annotation costs. (Longlong Jing and Tian 2019)

  • Consider implementing a Feature Fusion Learning (FFL) framework for efficient training of powerful classifiers. This involves creating a fusion module that combines feature maps from parallel neural networks, resulting in more meaningful feature maps. Additionally, the authors suggest incorporating an online mutual knowledge distillation system, wherein an ensemble of sub-network classifiers transfer your knowledge to the fused classifier, and vice versa. This mutual teaching system not only improves the performance of the (Jangho Kim et al. 2019)

  • Consider utilizing the HyperNOMAD package, which employs the Mesh Adaptive Direct Search (MADS) algorithm, to efficiently optimize the hyperparameters of deep neural networks, thereby improving your performance and reducing the time spent on manual tuning. (Lakhmiri, Digabel, and Tribes 2019)

  • Utilize the “Smoothly Varying Weight Hypothesis” (SVWH) in your deep neural network designs. This hypothesis suggests that the weights in adjacent convolution layers share strong similarity in shapes and values, allowing for more effective compression and quantization of the predicted residuals between the weights in all or adjacent convolution layers. By doing so, researchers can achieve a higher weight compression rate at the same accuracy level compared to previous quantization-based compression methods in deep neural networks (K.-H. Lee, Jeong, and Bae 2019)

  • Consider using a novel network pruning technique that generates a low-rank binary index matrix to compress index data while decomposing index data is performed by simple binary matrix multiplication, resulting in improved efficiency and reduced memory footprint. (D. Lee et al. 2019)

  • Consider incorporating a dynamic selection mechanism in your Convolutional Neural Networks (CNNs) designs, allowing each neuron to adaptively adjust its receptive field size based on multiple scales of input information. This can lead to improved performance and reduced model complexity. (Xiang Li et al. 2019)

  • Consider incorporating a neural-symbolic capsule architecture into your studies, particularly when dealing with inverse graphics problems. This architecture combines the strengths of neural networks and symbolic reasoning, enabling better understanding and manipulation of complex scenes through continuous improvement via lifelong meta-learning. (M.-Y. Liu et al. 2019)

  • Consider using high-level synthesis (HLS) tools like Xilinxs SDSoC to simplify the design and deployment of FPGA accelerators for deep learning applications, even within complex FPGA systems-on-chips (SoCs). (Mousouliotis and Petrou 2019)

  • Consider using a bounded variant of the L1 regularizer to achieve higher pruning rates and maintain generalization performance in deep neural networks. (Mummadi et al. 2019)

  • Consider utilizing the proposed hyperbolic wrapped distribution for gradient-based learning in probabilistic models on hyperbolic space, enabling efficient sampling and avoidance of auxiliary methods like rejection sampling. (Nagano et al. 2019)

  • Consider using a differentiable search space that allows for annealing of architecture weights and gradual pruning of inferior operations to improve the efficiency and accuracy of neural architecture searches. (Noy et al. 2019)

  • Consider using feature-level ensemble for knowledge distillation (FEED) to effectively transfer knowledge from multiple teacher networks to a student network, improving overall performance without increasing computational costs. (S. Park and Kwak 2019)

  • Focus on creating a balance between speed and ease of use in your designs, while also considering the importance of interoperability and extensibility within the Python ecosystem. (Paszke et al. 2019)

  • Separate and optimize convolutional and fully connected layers individually within deep neural networks to enhance your performance. (B. Qian and Wang 2019)

  • Carefully consider the choice of training hyper-parameters when applying theory-trained neural networks to solve partial differential equations, as this can greatly impact the success and efficiency of the training process. (Rad et al. 2019)

  • Utilize a differentiable mask when pruning convolutional and recurrent networks, allowing for greater sparsity and improved performance. (Ramakrishnan, Sari, and Nia 2019)

  • Consider utilizing spectral-domain Generative Adversarial Networks (GANs) when dealing with high-resolution 3D point-cloud generation tasks, as this approach simplifies the learning task and enables the production of high-quality point-clouds with minimal computational overhead. (Ramasinghe et al. 2019)

  • Focus on developing techniques that enable deep neural networks to efficiently utilize available hardware resources, specifically by employing structured pruning methods that promote parallelism and reduce memory usage. (Schindler et al. 2019)

  • Consider revising your neural networks to incorporate rotation-equivariant quaternion neural networks (REQNNs) for better handling of 3D point cloud processing tasks, as they provide both rotation equivariance and permutation invariance properties. (W. Shen et al. 2019)

  • Consider combining knowledge distillation and quantization techniques to effectively compress acoustic event detection models, resulting in reduced error rates and model sizes suitable for deployment on devices with limited computational resources. (B. Shi et al. 2019)

  • Focus on selecting appropriate temperature values for the softmax distribution in order to optimize the performance of quantized deep neural networks through knowledge distillation techniques. (S. Shin, Boo, and Sung 2019)

  • Consider integrating hierarchical clustering techniques into your representation learning models to better capture the underlying structure of complex data. (S.-J. Shin, Song, and Moon 2019)

  • Consider utilizing a novel ensemble approach for embedding distillation in order to improve the efficiency and accuracy of deep neural models in NLP tasks. (B. Shin, Yang, and Choi 2019)

  • Employ a tree-structured graph convolution network (TreeGCN) as a generator for tree-GAN when aiming to achieve state-of-the-art performance for multi-class 3D point cloud generation. (Shu, Park, and Kwon 2019)

  • Consider combining convolutions and attention mechanisms in your neural network architectures to leverage the strengths of both layer types, while also exploring efficient search strategies like Progressive Dynamic Hurdles to identify optimal architectures within large search spaces. (So, Liang, and Le 2019)

  • Consider using the proposed modification to the loss function (Equation 1) to eliminate all bad local minima from any loss landscape, without requiring additional units or assumptions about the nature of the loss. (Sohl-Dickstein and Kawaguchi 2019)

  • Utilise a more comprehensive and varied dataset like Meta-Dataset for few-shot classification tasks, rather than relying solely on limited datasets such as Omniglot and mini-ImageNet. (Triantafillou et al. 2019)

  • Consider developing an automated compiler-based FPGA accelerator for efficient and scalable training of convolutional neural networks (CNNs) across various architectural configurations. (Venkataramanaiah et al. 2019)

  • Utilise a multiscale visualisation tool to better understand and interpret the complex attention mechanisms in transformer models, allowing for improved model transparency and facilitation of various applications such as detecting model biases, locating relevant attention heads, and linking neurons to model behaviour. (Vig 2019)

  • Explore the potential impact of universal adversarial triggers on various NLP models, as they can reveal critical vulnerabilities and offer valuable insights into global model behavior. (Wallace et al. 2019)

  • Integrate the ranking phase and the fine-tuning phase by sharing intermediate computation results in order to significantly reduce the ranking time while maintaining high classification accuracy. (Zi Wang et al. 2019)

  • Consider using graph convolution networks (GCNs) to improve the accuracy and scalability of face clustering tasks, particularly when dealing with complex distributions of face representations. (Zhongdao Wang et al. 2019)

  • Consider using structured pruning methods for reducing the overall storage and computation costs of recurrent neural networks (RNNs) by selecting independent neurons, rather than relying solely on traditional Lasso-based pruning methods that produce irregular sparse patterns in weight matrices. (L. Wen et al. 2019)

  • Consider using continuous normalizing flows to model the distribution of points given a shape, enabling accurate sampling and estimation of probability densities within a principled probabilistic framework. (Guandao Yang et al. 2019)

  • Consider implementing a multi-task knowledge distillation model (MKDM) for model compression, which involves training multiple teacher models to obtain knowledge and then designing a multi-task framework to train a single student model by leveraging multiple teachers knowledge, thereby improving generalization performance and reducing over-fitting bias during the distillation stage.’ (Z. Yang et al. 2019)

  • Consider applying deep model quantization and compression to your Convolutional Neural Network (CNN) models when working with low-power hardware implementations, such as ASIC engines, for tasks like image retrieval. (Bin Yang et al. 2019)

  • Focus on developing a transformer-based framework called Prompt Promotion’, which uses metapath- and embedding-based prompts to enhance the model’s predictions for undetermined connection patterns in the app promotion graph. (L. Yao, Mao, and Luo 2019)

  • Consider using Teacher-free Knowledge Distillation (Tf-KD) instead of traditional Knowledge Distillation (KD) methods, as Tf-KD allows for comparable performance improvements without requiring a separate teacher model, making it particularly useful in situations where finding a suitable teacher model is difficult or computationally expensive. (Li Yuan et al. 2019)

  • Consider using the Incremental Pruning Based on Less Training (IPLT) algorithm, which reduces the amount of pre-training required for pruning algorithms, resulting in faster and more efficient model compression. (Yue, Weibin, and Lin 2019)

  • Integrate deep learning techniques with existing mathematical models to create a hybrid approach for designing and optimizing future wireless communication networks. (Zappone, Renzo, and Debbah 2019)

  • Continually develop and break datasets in order to create dynamic benchmarks that evolve alongside advances in artificial intelligence technology. (Zellers, Holtzman, Bisk, et al. 2019)

  • Consider incorporating rotation invariant geometric features such as distances and angles into your convolution operators for point cloud learning, as this can improve the overall robustness and generalizability of your models. (Zhiyuan Zhang et al. 2019)

  • Focus on developing novel operators, such as Graph Embedding Module (GEM) and Pyramid Attention Network (PAN), to effectively capture local geometric relationships and improve the overall performance of point cloud classification and semantic segmentation tasks. (Zhiheng and Ning 2019)

  • Use sinusoidal mapping of inputs in g-PINN architectures to increase input gradient variability, thereby avoiding getting trapped in deceptive local minima caused by initial biases towards flat output functions in physics-informed neural networks. (M. Raissi, Perdikaris, and Karniadakis 2019)

  • Employ Finite Basis Physics-Informed Neural Networks (FBPINNs) to overcome the limitations of conventional Physics Informed Neural Networks (PINNs) in solving large-scale differential equation problems. (Giorgi 2019)

  • Focus on understanding the balance between innate and learned behaviors in animals, as well as exploring the potential benefits of incorporating innate mechanisms into artificial neural networks to improve your efficiency and effectiveness. (Zador 2019)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Price, Bethune, and Massey 2019)

  • Consider combining the strengths of fuzzing and symbolic execution by learning a fuzzer from inputs generated by a symbolic execution expert using the framework of imitation learning, resulting in a faster and more effective way to generate test inputs for software testing. (J. He et al. 2019)

  • Consider adopting an algorithm-hardware co-design approach when developing Convolutional Neural Network (ConvNet) accelerators for Field Programmable Gate Arrays (FPGA). This involves creating a ConvNet model specifically tailored to FPGA requirements, like the DiracDeltaNet model introduced in the study, which enables the creation of a highly customised computing unit for the FPGA. (Yifan Yang et al. 2019)

  • Consider combining various machine learning approaches, such as deep neural networks, gradient boosted decision trees, and factorization machines, to achieve optimal results in complex tasks like search ranking. (Haldar et al. 2019)

  • Utilise Collaborative Knowledge Graphs (CKGs) when making recommendations. These graphs combine user behaviour and item knowledge into a unified relational graph, allowing for better understanding of user preferences and improved recommendation accuracy. (Xiang Wang et al. 2019)

  • Consider using the DeepSZ framework for lossy compression of deep neural networks, which involves network pruning, error bound assessment, optimization of error bound configuration, and compressed model generation, resulting in improved compression ratios and reduced storage requirements while maintaining high inference accuracy. (S. Jin et al. 2019)

  • Carefully examine the relationship between the models choice of prices and what guests actually prefer, and ensure that the model takes into account the “cheaper is better” principle when ranking listings. (Aman Agarwal et al. 2019)

  • Use a combination of region-wise convolutions and non-local correlations within a coarse-to-fine framework to achieve better image inpainting results, particularly for large irregular missing regions. (Yuqing Ma et al. 2019)

  • Utilize both local and global anomaly detection methods when analyzing social media data to accurately identify rumors, as relying solely on either method could result in false positives or negatives. (Tam et al. 2019)

  • Utilise a neural network model to predict the latent naturalness score’ of ConceptNet paths based on crowdsource assessment data, instead of relying solely on heuristic methods. (Yilun Zhou, Schockaert, and Shah 2019)

  • Consider utilising the Nengo and Nengo_extras packages to convert Deep Neural Networks (DNNs) to Spiking Neural Networks (SNNs) and incorporate Permadrop layers within the Nengo framework to improve the efficiency and accuracy of your modelling efforts. (N. Baker et al. 2018)

  • Carefully consider the tradeoff between model simplicity and prediction accuracy when developing statistical models, particularly in situations where parsimony is desired. (A. Zhou et al. 2018)

  • Focus on developing specialized FPGA accelerators for specific deep convolutional neural network (DCNN) architectures, like SqueezeNet, to improve efficiency and reduce computational costs while maintaining high levels of accuracy in real-time applications. (Mousouliotis and Petrou 2018)

  • Utilize Capsule Networks (CapsNets) for brain tumor classification, as they offer advantages over traditional Convolutional Neural Networks (CNNs) in terms of requiring less training data, being more robust to rotation and affine transformation, and potentially offering better classification accuracy. (Afshar, Mohammadi, and Plataniotis 2018)

  • Consider exploiting the high locality inherent in large language model (LLM) inference, characterized by a power-law distribution in neuron activation, to optimize the efficiency of neuron activation and computational sparsity. (Agarap 2018)

  • Consider casting neural network quantization as a discrete labelling problem, and examine relaxations to develop an efficient iterative optimization procedure involving stochastic gradient descent followed by a projection, ultimately proving that your proposed simple projected gradient descent approach is equivalent to a proximal version of the well-established mean-field method. (Ajanthan et al. 2018)

  • Employ an intervention-based behavioural analysis paradigm to evaluate the behaviour of Vision-and-Language Navigation (VLN) agents. (P. Anderson et al. 2018)

  • Focus on developing methods that leverage noise stability properties of deep nets to achieve better compression and generalization performance. (Sanjeev Arora et al. 2018)

  • Consider using a Swapping Autoencoder for deep image manipulation, as it effectively disentangles texture from structure, allowing for accurate and realistic image reconstruction, while being substantially more efficient compared to recent generative models. (Asim, Shamshad, and Ahmed 2018)

  • Carefully select appropriate machine learning algorithms to optimize brain tumor segmentation, progression assessment, and overall survival prediction in the context of the BRATS challenge. (Bakas et al. 2018)

  • Consider implementing a novel 4-bit post-training quantization technique for convolutional neural networks, which combines three complementary methods for minimizing quantization error at the tensor level, leading to improved accuracy and reduced computational requirements. (Banner et al. 2018)

  • Utilise ensemble methods to reduce the variance of few-shot learning classifiers, thereby improving your overall performance. (Bietti et al. 2018)

  • Leverage the power of Contrastive Language-Image Pre-training (CLIP) models to develop a text-based interface for StyleGAN image manipulation, eliminating the need for manual effort or annotated collections of images for each desired manipulation. (Brock, Donahue, and Simonyan 2018)

  • Consider adopting a machine learning-based approach to jointly optimize both neural and hardware architecture, leading to significant improvements in speed and energy savings without compromising accuracy. (Han Cai, Zhu, and Han 2018)

  • Consider using Knowledge Distillation with Feature Maps (KDFM) to improve the efficiency of deep learning models while maintaining accuracy, particularly for image classification tasks. (W.-C. Chen et al. 2018)

  • Focus on developing data-free network compression methods like PNMQ, which employ Parametric Non-uniform Mixed Precision Quantization to efficiently compress deep neural networks while preserving your quality, without requiring extensive datasets or costly computations. (Zhuo Chen et al. 2018)

  • Employ a Progressive Feature Alignment Network (PFAN) for effective unsupervised domain adaptation (UDA), which involves an Easy-to-Hard Transfer Strategy (EHTS) and an Adaptive Prototype Alignment (APA) step to train the model iteratively and alternatively, ensuring cross-domain category consistency and reducing error accumulation. (Chaoqi Chen et al. 2018)

  • Utilize a deep reinforcement learning framework called ReLeQ to automate the discovery of optimal quantization levels for deep neural networks, thereby balancing speed and quality while preserving accuracy and reducing computational and storage costs. (Elthakeb et al. 2018)

  • Utilise hypergraph neural networks (HGNN) for data representation learning, particularly when dealing with complex and high-order data correlations. (Yifan Feng et al. 2018)

  • Utilize a novel deep architecture that learns topologically interpretable discrete representations in a probabilistic fashion, allowing for improved clustering and interpretability of time series data. (Fortuin et al. 2018)

  • Carefully examine and exploit input and kernel similarities in BNNs to significantly reduce computation redundancies and enhance the efficiency and speed of your inference processes. (Cheng Fu et al. 2018)

  • Consider adopting hyperbolic neural networks for handling complex data, particularly those with hierarchical or tree-like structures, as they offer superior performance compared to traditional Euclidean embeddings. (Ganea, Bécigneul, and Hofmann 2018)

  • Focus on creating a few-shot visual learning system that can effectively learn novel categories from limited training data while preserving the original categories information, thereby improving overall recognition performance.’ (Gidaris and Komodakis 2018)

  • Consider implementing a novel deep neural network training technique called Dropback, which reduces the number of weights updated during backpropagation to those with the highest total gradients, thereby significantly decreasing the number of off-chip memory accesses during both training and inference, leading to potential improvements in energy efficiency and accuracy retention. (Golub, Lemieux, and Lis 2018)

  • Utilize retrieval-based techniques for prompt selection in order to effectively demonstrate code-related tasks in few-shot learning scenarios. (Hata, Shihab, and Neubig 2018)

  • Utilize a full variational distribution over weights instead of deterministic weights, allowing for more efficient coding schemes and higher compression rates in deep neural networks. (Havasi, Peharz, and Hernández-Lobato 2018)

  • Utilise statistical weight scaling and residual expansion methods to reduce the bit-width of the whole network weight parameters to ternary values, thereby reducing model size, computation cost, and minimising accuracy degradation caused by model compression. (Zhezhi He, Gong, and Fan 2018)

  • Consider employing model-driven deep learning techniques in physical layer communications, as they provide a balance between leveraging domain knowledge and harnessing the power of deep learning, leading to lower data requirements, reduced risk of overfitting, and quicker implementation. (Hengtao He et al. 2018)

  • Consider the explicit impact of ternarization on the loss function when developing weight ternarization techniques for deep neural networks, and optimize accordingly. (L. Hou and Kwok 2018)

  • Leverage stochastic optimization techniques in the pruning process of deep neural networks to avoid deleting globally important weights and allow them to potentially return, thereby improving overall model compression and accuracy performance. (H. Jia et al. 2018)

  • Utilize a style-based generator architecture for generative adversarial networks, which borrows from style transfer literature, to achieve an automatically learned, unsupervised separation of high-level attributes and stochastic variation in generated images, resulting in improved performance across traditional distribution quality metrics, better interpolation properties, and superior disentangling of latent factors of variation. (Karras, Laine, and Aila 2018)

  • Consider implementing a neural network-hardware co-design approach to optimize the performance of RRAM-based BNN accelerators by splitting input data to fit each split network on a RRAM array, allowing for 1-bit output neuron calculations in each array and eliminating the need for high-resolution ADCs. (Yulhwa Kim, Kim, and Kim 2018)

  • Consider using FactorVAE, a novel method that provides a better balance between disentanglement and reconstruction quality compared to existing techniques, such as beta-VAE, for unsupervised learning of disentangled representations. (Hyunjik Kim and Mnih 2018)

  • Consider utilizing a novel knowledge transfer method involving convolutional operations to paraphrase teachers knowledge and translate it for the student, resulting in improved performance of the student network.’ (Jangho Kim, Park, and Kwak 2018)

  • Utilize a novel method for compute-constrained structured channel-wise pruning of convolutional neural networks, which involves iteratively fine-tuning the network while gradually tapering the computation resources available to the pruned network via a holonomic constraint in the method of Lagrangian multipliers framework. (Kruglov 2018)

  • Utilise a combination of metric learning and adversarial learning techniques for effective unsupervised domain adaptation, leading to significant improvements in classification accuracy. (Laradji and Babanezhad 2018)

  • Utilise the Knowledge Distillation’ technique to convert complex Deep Neural Networks into simpler, more interpretable decision trees. This allows for improved understanding and reasoning behind the predictions, making the models more transparent and trustworthy, particularly in areas where ethics and mission-critical applications are involved. (Xuan Liu, Wang, and Matwin 2018)

  • Consider combining channel pruning and model fine-tuning into a single end-to-end trainable system for improved results in deep model inference efficiency. (J.-H. Luo and Wu 2018)

  • Carefully consider the choice of transliteration method, as well as the quality and quantity of training data, when developing a multilingual named entity transliteration system. (Merhav and Ash 2018)

  • Focus on developing a novel representation for 3D geometry based on learning a continuous 3D mapping, which can be used for reconstructing 3D geometry from various input types and generates high-quality meshes. (Mescheder et al. 2018)

  • Consider integrating language information into meta-learning algorithms to enhance the efficiency and adaptability of artificial agents when interacting with novel tools. (Nichol, Achiam, and Schulman 2018)

  • Consider employing a technique called Deep Net Triage’, which involves systematically compressing, initialising, and training neural network layers to determine your criticality and impact on overall network performance.’ (Nowak and Corso 2018)

  • Consider combining multiple methods of model compression, such as pruning and knowledge distillation, to achieve significantly reduced model sizes while maintaining high levels of accuracy. (Oguntola, Olubeko, and Sweeney 2018)

  • Consider employing Universal Differential Equations (UDEs) as a novel methodology for combining mechanistic models and data-driven machine learning approaches, allowing them to leverage the strengths of both while addressing your respective limitations. (Otter, Medina, and Kalita 2018)

  • Adopt a distribution-aware approach to binarizing deep neural networks, allowing them to maintain the advantages of a binarized network while reducing accuracy drops. (Prabhu et al. 2018)

  • Focus on understanding filter functionality when conducting filter pruning in Convolutional Neural Networks (CNNs), instead of solely relying on filter magnitude ranking methods like (_{1}) norm, to avoid compromising the overall network performance. (Zhuwei Qin et al. 2018)

  • Adopt a novel feature extraction model based on a sparse autoencoder within a bag-of-features framework for text recognition, followed by utilizing hidden markov models for sequencing. (Rahal, Tounsi, and Alimi 2018)

  • Prioritize the development of neural network-based models for estimating the likelihood of two-way interest between candidates and recruiters, and the learning of supervised and unsupervised embeddings of entities in the talent search domain. (Ramanath et al. 2018)

  • Carefully consider the impact of both application-level specifications (such as neural network data, layers, and activation functions) and architectural-level specifications (like data representation model and parallelism degree of the underlying accelerator) when studying the resilience of RTL NN accelerators. (Salami, Unsal, and Cristal 2018)

  • Utilise a hierarchical multi-task approach for learning embeddings from semantic tasks, which involves training a model in a hierarchical manner to introduce an inductive bias by supervising a set of low level tasks at the bottom layers of the model and more complex tasks at the top layers of the model. (Sanh, Wolf, and Ruder 2018)

  • Carefully consider the actual SNN operation during the ANN-SNN conversion process, as demonstrated by the proposed weight-normalization technique that accounts for the actual SNN operation, leading to near-lossless ANN-SNN conversion for significantly deep architectures and complex recognition problems. (Sengupta et al. 2018)

  • Consider using a subtractive definition of prosody, which involves accounting for variations due to phonetics, speaker identity, and channel effects before analyzing the remaining variation in speech signals. (Skerry-Ryan et al. 2018)

  • Carefully consider the choice of compression method for deep neural networks, as the authors demonstrate that your novel DeepThin technique outperforms several existing methods in terms of accuracy and compression rate. (Sotoudeh and Baghsorkhi 2018)

  • Use tensorial neural networks (TNNs) instead of traditional neural networks (NNs) because TNNs offer superior flexibility and expressivity, enabling them to capture multidimensional structures in the input data and improve model compression. (Jiahao Su et al. 2018)

  • Consider using Principal Filter Analysis (PFA) for neural network compression, as it effectively reduces network size while preserving accuracy through analyzing the correlation within the responses of each layer. (Suau, Zappella, and Apostoloff 2018)

  • Consider using the MPDCompress algorithm when working with deep neural networks (DNNs) to effectively compress the network without compromising its accuracy, making it suitable for deployment on edge devices in real-time. (Supic et al. 2018)

  • Consider employing deep transfer learning strategies to overcome the challenge of insufficient training data in certain domains, such as bioinformatics and robotics, by leveraging knowledge from other domains through deep neural networks. (Chuanqi Tan et al. 2018)

  • Focus on developing entropy-based unsupervised domain adaptation strategies for improving semantic segmentation performance in various scenarios, especially those involving synthetic-to-real transitions. (Vu et al. 2018)

  • Consider utilizing a hardware-aware automated quantization (HAQ) framework that incorporates reinforcement learning to intelligently allocate bitwidths for weights and activations across different layers of a neural network, thereby optimizing latency, energy consumption, and storage on target hardware without requiring domain experts or rule-based heuristics. (Kuan Wang et al. 2018)

  • Consider developing a novel method called “WAGE” to discretize both training and inference processes in deep neural networks, allowing for improved accuracies and potentially enabling deployment in hardware systems such as integer-based deep learning accelerators and neuromorphic chips. (S. Wu et al. 2018)

  • Utilise alternating minimisation to effectively quantify recurrent neural networks, resulting in significant improvements in memory savings and real inference acceleration without compromising accuracy. (Chen Xu et al. 2018)

  • Use attention statistics, a novel attention-based criterion for channel pruning, to optimize the appended neural networks and enable accurate estimation of redundant channels, thereby achieving superior performance over conventional methods in terms of accuracy and computational costs for various models and datasets. (K. Yamamoto and Maeno 2018)

  • Utilise snapshot distillation (SD) for teacher-student optimization in one generation, which significantly reduces computational overheads and enhances the overall performance of deep neural networks. (Chenglin Yang et al. 2018)

  • Consider using a bilinear regression model to estimate the energy consumption of deep neural networks (DNNs) when developing energy-constrained DNN compression frameworks. (Haichuan Yang, Zhu, and Liu 2018a)

  • Incorporate energy constraints into deep neural network training processes, allowing for efficient optimization and improved accuracy under specified energy budgets. (Haichuan Yang, Zhu, and Liu 2018b)

  • Utilise the Alternating Direction Method of Multipliers (ADMM) as a unifying approach to tackle complex, non-convex optimization problems in deep neural networks (DNNs), specifically those involving weight pruning and clustering/quantization. (S. Ye et al. 2018)

  • Consider implementing Self-Attention Generative Adversarial Networks (SAGANs) for image synthesis tasks, as they allow for efficient modeling of long-range dependencies and improve overall performance. (Han Zhang et al. 2018)

  • Consider incorporating a Variational Autoencoder (VAE) module into your end-to-end Text-to-Speech (TTS) model to enable unsupervised learning of the latent representation of speaking styles, thereby facilitating effective style control and transfer in synthesized speech. (Y.-J. Zhang et al. 2018)

  • Adopt a novel approach to interpreting neural networks by partitioning the space of sequences of neuron activations, leading to improved understanding and control over complex models. (Zharov et al. 2018)

  • Use a neural pattern diagnosis framework like DIAG-NRE to automatically summarize and refine high-quality relational patterns from noise data with human experts in the loop, thereby improving the efficiency and generalizability of distantly supervised neural relation extraction. (S. Zheng et al. 2018)

  • Utilise deep convolutional neural networks (DCNNs) due to your proven universality, allowing them to approximate any continuous function to an arbitrary accuracy when the depth of the neural network is large enough. (D.-X. Zhou 2018)

  • Utilise path-based abstractions of a programs abstract syntax tree (AST) to create a general, fully automatic, and cross-language compatible representation of source code for learning purposes.’ (Yahav 2018)

  • Consider using the Quasi-Lloyd-Max algorithm to minimize weight quantization error when working with 4-bit networks, leading to improved accuracy and reduced fine-tuning time. (Jian Cheng et al. 2018)

  • Consider utilizing a hybrid deep learning approach combining convolutional neural networks (CNN) and bi-directional short term memory (BDLSTM) networks to effectively recognize Arabic text in images, even those with varying font types and cursive styles. (Alghamdi and Teahan 2018)

  • Focus on developing methods to optimize the implementation of binarized neural networks (BNNs) on field programmable gate arrays (FPGAs) using techniques such as resource-aware model analysis (RAMA), datapath design with XNOR, popcount, and shifting operations, and optimized data management strategies to achieve high performance and energy efficiency. (S. Liang et al. 2018)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Speer and Lowry-Duda 2018)

  • Carefully examine the effects of quantization techniques on individual layers of a neural network, taking into account the range of data, precision of variables, and position of the layer within the network, in order to optimize memory usage and computational speed without sacrificing accuracy. (Prado et al. 2018)

  • Focus on developing methods that can effectively capture the heterogeneity of field pair interactions in multi-field categorical data, leading to improved predictive performance and reduced model complexity. (Junwei Pan et al. 2018)

  • Utilise deep learning algorithms, specifically graph neural networks, to effectively learn users latent feature representations for accurate social influence predictions across diverse social media platforms.’ (J. Qiu et al. 2018)

  • Carefully consider the tradeoff between effectiveness and efficiency in developing ranking models, and explore techniques such as ranking distillation to improve both aspects simultaneously. (Jiaxi Tang and Wang 2018)

  • Consider using fixed integer inter-layer signals and fixed-point weights in order to maintain good accuracy while reducing the need for extensive data computations in deep neural networks. (F. Liu and Liu 2018)

  • Leverage a large collection of actual review manipulators, rather than simulating or assuming the existence of fake reviews, in order to better understand and combat review manipulation in online systems. (Kaghazgaran, Caverlee, and Squicciarini 2018)

  • Focus on developing deep and wide neural networks like DAWnet to enhance the relevance, depth, and breadth of chatbot responses in multi-turn dialogue systems. (Wenjie Wang et al. 2018)

  • Consider using the Vector Quantized-Variational AutoEncoder (VQ-VAE) model for learning discrete representations without supervision, as it addresses the “posterior collapse” issue commonly encountered in Variational AutoEncoder (VAE) frameworks and generates high-quality images, videos, and speech. (Agustsson et al. 2017)

  • Utilise deep learning techniques to create optimal weighting systems for covariate balance in causal inference studies, thereby reducing bias and improving accuracy. (Arjovsky, Chintala, and Bottou 2017)

  • Focus on understanding the underlying mechanisms of existing algorithms rather than solely creating new ones, while also considering alternative approaches to traditional reinforcement learning frameworks. (Sanjeev Arora et al. 2017)

  • Focus on developing a compression framework for understanding generalization in deep neural networks, which involves identifying noise stability properties within the network and utilizing these properties to create efficient and provably correct algorithms for reducing the effective number of parameters in the network. (Arpit et al. 2017)

  • Focus on developing alternative approaches to uniform convergence for explaining generalization in deep learning, as current bounds derived from uniform convergence either grow with parameter count or require modification to the network. (Yoshua Bengio 2017)

  • Aim to obtain a certified and non-trivial lower bound on the minimum adversarial distortion for deep neural networks, ideally within a reasonable amount of computational time. (Carlini and Wagner 2017)

  • Consider developing universal architectures for image segmentation tasks, rather than focusing solely on specialized architectures, as demonstrated by the Masked-attention Mask Transformer (Mask2Former) which outperforms specialized architectures across various segmentation tasks while remaining easy to train. (L.-C. Chen et al. 2017)

  • Consider utilising a combination of temporal convolutional neural networks (TCNNs) and transfer learning to enhance the efficiency and effectiveness of video classification tasks. (Diba et al. 2017)

  • Utilize a nonstationary multi-armed bandit algorithm to optimize learning progress in neural networks, based on a reward signal derived from the rate of increase in prediction accuracy or network complexity. (Graves et al. 2017)

  • Consider modifying your training regime to include a higher learning rate and batch normalization, as this approach can help close the generalization gap in large batch training of neural networks. (Hoffer, Hubara, and Soudry 2017)

  • Incorporate the “Spatio-Temporal Channel Correlation” (STC) block into your 3D CNN architectures to enhance the performance of action classification tasks by effectively modelling correlations between channels of a 3D CNN with respect to temporal and spatial features. (Jie Hu et al. 2017)

  • Utilize the TriviaQA dataset, which features complex, compositional questions with considerable syntactic and lexical variability, and necessitates cross-sentence reasoning to locate answers, thus providing a robust testing ground for reading comprehension models. (M. Joshi et al. 2017)

  • Consider using cosine normalization, which replaces the traditional dot product calculation in neural networks with cosine similarity, leading to improved stability and reduced variance compared to other normalization techniques such as batch, weight, and layer normalization. (Chunjie Luo et al. 2017)

  • Consider employing a combination of evolutionary optimization processes at different levels to optimize the design of deep neural networks, allowing for efficient exploration of a wider range of potential solutions. (Miikkulainen et al. 2017)

  • Consider implementing virtual adversarial training (VAT) as a regularization method for supervised and semi-supervised learning tasks, as it effectively addresses the issue of overfitting by promoting local distributional smoothness (LDS) through an efficient approximation of the virtual adversarial loss, leading to improved generalization performance across multiple benchmark datasets. (Miyato et al. 2017)

  • Consider adopting the dynamic declaration programming model for implementing neural network models, as it enables greater flexibility in handling complex network architectures and simplifies the implementation process compared to traditional static declaration strategies. (Neubig et al. 2017)

  • Consider using Probability Density Distillation when working with autoregressive models like WaveNet, as it enables efficient training and accurate prediction of high-quality speech samples. (Oord et al. 2017)

  • Consider using the Vector Quantized-Variational AutoEncoder (VQ-VAE) model for learning discrete representations in machine learning, as it addresses the “posterior collapse” issue commonly encountered in Variational AutoEncoder (VAE) frameworks and provides high-quality samples across various applications. (Oord, Vinyals, and Kavukcuoglu 2017)

  • Carefully choose auxiliary tasks that complement your primary task, allowing them to leverage the benefits of multi-task learning in deep neural networks, including improved generalization, reduced overfitting, and increased sample efficiency. (Ruder 2017)

  • Consider adopting a “temporal segment network” (TSN) framework for action recognition tasks in videos. This involves using a sparse and global temporal sampling strategy to efficiently model long-range temporal structures across the entire video, rather than focusing solely on appearances and short-term motions. (Limin Wang et al. 2017)

  • Utilise the DeepSets architecture when dealing with machine learning tasks involving sets, as it allows for permutation invariant and equivariant functions, enabling accurate predictions regardless of the order of elements in the set. (Zaheer et al. 2017)

  • Utilize deep learning-based numerical methods for solving high-dimensional parabolic partial differential equations (PDEs) and backward stochastic differential equations (BSDEs) by leveraging the analogy between BSDEs and reinforcement learning, where the gradient of the solution acts as the policy function and the loss function represents the difference between the prescribed terminal condition and the BSDE solution. (E, Han, and Jentzen 2017)

  • Develop a Spatial Incomplete Multi-task Deep leArning (SIMDA) framework for effective forecasting of spatio-temporal event subtypes, incorporating spatial heterogeneity, incomplete labeling, and profound representations of event subtypes. (“Open Source Indicators Project” 2017)

  • Employ a tree-structured graph convolution network (TreeGCN) as a generator for tree-GAN when aiming to achieve state-of-the-art performance for multi-class 3D point cloud generation., ‘This paper emphasizes the importance of using a tree-structured graph convolution network (TreeGCN) as a generator for tree-GAN to attain superior performance in multi-class 3D point cloud generation.’ (Arjovsky, Chintala, and Bottou 2017)

  • Consider incorporating Gaussian processes (GPs) within deep neural networks (DNNs) to improve uncertainty estimation and enhance robustness against adversarial examples. (Bradshaw, G. Matthews, and Ghahramani 2017)

  • Utilise the learning-compression’ (LC) algorithm when dealing with neural network quantisation. This algorithm provides a clear separation between learning and quantification, allowing for easier computational processes and ensuring that the final output is a valid solution.’ (Carreira-Perpiñán and Idelbayev 2017)

  • Carefully consider the trade-offs between model size and retrieval performance when developing compressed deep neural networks for image instance retrieval tasks, utilizing techniques such as quantization, coding, pruning, and weight sharing. (Chandrasekhar et al. 2017)

  • Consider utilizing a GAN inversion process when attempting to solve the image outpainting problem, as it allows for the discovery of multiple latent codes that produce diverse outpainted regions, ultimately resulting in increased diversity and richness in the outpainted areas. (L.-C. Chen et al. 2017)

  • Focus on developing efficient and accurate student networks by leveraging the benefits of structural model distillation, specifically through attention transfer, to achieve significant memory savings with minimal loss of accuracy. (Crowley, Gray, and Storkey 2017)

  • Consider using a combination of adversarial and L1 losses when training GANs for speech enhancement, as it leads to better performance compared to using just the adversarial loss. (C. Donahue, Li, and Prabhavalkar 2017)

  • Consider using a compound scaling method to uniformly scale network width, depth, and resolution in a principled manner, leading to improved accuracy and efficiency in Convolutional Neural Networks. (Howard et al. 2017)

  • Consider using the Maximum Mean Discrepancy (MMD) metric to minimize the difference in neuron selectivity patterns between teacher and student networks during knowledge transfer processes. (Zehao Huang and Wang 2017b)

  • Apply Sparse Variational Dropout to linear models to achieve a sparse solution while providing the Automatic Relevance Determination effect, thereby overcoming certain disadvantages of empirical Bayes. (D. Molchanov, Ashukha, and Vetrov 2017)

  • Use dynamic estimation of quantization step sizes during retraining to improve the performance of fixed-point optimization of deep neural networks. (S. Shin, Boo, and Sung 2017)

  • Consider using prototypical networks for few-shot and zero-shot learning problems, as they offer a simpler inductive bias and achieve excellent results compared to recent approaches involving complex architectural choices and meta-learning. (Snell, Swersky, and Zemel 2017)

  • Adopt a “fine-pruning” approach when working with pre-trained convolutional networks, which combines fine-tuning and compression into a single iterative process, thereby improving the overall efficiency and effectiveness of the network. (Tung, Muralidharan, and Mori 2017)

  • Utilize the fpgaConvNet framework to map diverse Convolutional Neural Networks onto embedded FPGAs using an automated design methodology based on the Synchronous Dataflow (SDF) paradigm, allowing for efficient exploration of the architectural design space and generation of optimized hardware designs for various performance metrics. (Venieris and Bouganis 2017)

  • Consider implementing a deep mutual learning (DML) strategy, where instead of one-way transfer between a static pre-defined teacher and a student, an ensemble of students learn collaboratively and teach each other throughout the training process, leading to improved performance on tasks like CIFAR-100 recognition and Market-1501 person re-identification. (Ying Zhang et al. 2017)

  • Utilise a novel post-training quantisation (PTQ) scheme named “subset quantization” (SQ) to improve the performance of your deep neural networks (DNNs) without increasing hardware costs. (Y.-H. Chen et al. 2017)

  • Carefully consider the potential impact of systematic diffusion when combining label smoothing and knowledge distillation techniques, as it could lead to reduced effectiveness of the distillation process. (Chorowski and Jaitly 2017)

  • Utilise the TensorDiffEq tool, which offers a scalable, modular, and customisable multi-GPU architecture and solver for Physics-Informed Neural Networks (PINNs), enabling efficient and accurate solutions for complex scientific problems. (Rackauckas and Nie 2017)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Canchola 2017)

  • Carefully evaluate the trade-off between compression factor, accuracy, and runtime when choosing a compression technique for recurrent neural networks, and that the proposed Hybrid Matrix Decomposition (HMD) approach offers a balance between these factors. (C. Ding et al. 2017)

  • Consider implementing a fusion architecture that combines multiple layers of a convolutional neural network (CNN) to reduce memory bandwidth requirements and increase overall efficiency. (Q. Xiao et al. 2017)

  • Consider incorporating an attention-based neural model that looks “in-between” rather than “across”, allowing them to explicitly model contrast and incongruity in sarcasm detection tasks. (Peled and Reichart 2017)

  • Utilize Elastic Weight Consolidation (EWC) to prevent catastrophic forgetting in neural networks by selectively slowing down learning on the weights important for previously learned tasks. (Kirkpatrick et al. 2017)

  • Utilize the dynr’ package for efficiently analyzing intensive longitudinal datasets with complex dynamics, including regime-switching properties, through a combination of computational efficiency and user-friendly model specification functions.’ (Pritikin, Rappaport, and Neale 2017)

  • Develop visualization techniques for recurrent neural networks (RNNs) that are easily interpretable by non-experts, allowing for better understanding and trust in AI systems. (Goodman and Flaxman 2017)

  • Consider using a hierarchical Gaussian mixture model (hGMM) when working with point clouds, as it allows for coarse-to-fine learning and consistent partitioning of the input shape, leading to improved results in tasks such as shape generation and registration. (Achlioptas et al. 2017)

  • Avoid treating attention weights as direct indicators of feature importance or as unique explanations for model predictions, since they often fail to correlate strongly with gradient-based measures of feature importance and can be replaced by alternative attention distributions that yield equivalent predictions. (Alvarez-Melis and Jaakkola 2017)

  • Consider incorporating both bottom-up and top-down attention mechanisms in your studies, as doing so allows for better integration of visual and linguistic information, leading to improved performance in tasks like image captioning and visual question answering. (P. Anderson et al. 2017)

  • Utilize a semantic representation learning module to improve the performance of adversarial adaptation methods in unsupervised domain adaptation tasks. (Arjovsky and Bottou 2017)

  • Use the Earth Mover (EM) distance instead of other probability distances and divergences when measuring the similarity between model and real distributions, as it has better convergence properties and is more suitable for learning distributions supported by low dimensional manifolds. (Arjovsky, Chintala, and Bottou 2017)

  • Consider utilising a two-stage reinforcement learning approach when attempting to reduce the complexity of a neural network without compromising its performance. (Ashok et al. 2017)

  • Focus on creating open-source neural machine translation (NMT) toolkits that prioritize efficiency, modularity, and extensibility, allowing for the exploration of diverse model architectures, feature representations, and source modalities, while still delivering competitive performance and manageable training requirements. (Britz et al. 2017)

  • Consider implementing early stopping methods for hyperparameter optimization and architecture search using performance prediction models, which can lead to significant speedups in both processes. (Brock et al. 2017)

  • Focus on developing efficient and accurate forward and backward approximation functions for the ReLU activation function in deep neural networks, taking advantage of the statistics of network activations and batch normalization operations commonly used in the literature. (Z. Cai et al. 2017)

  • Utilise a learning-compression’ (LC) algorithm when pruning neural networks. This algorithm alternates between a ‘learning’ phase that optimises a regularised, data-dependent loss, and a ‘compression’ phase that marks weights for pruning in a data-independent manner. By doing so, the algorithm allows for automatic determination of the ideal number of weights to prune in each layer of the network, thereby preventing overfitting and improving overall efficiency (Carreira-Perpiñán and Idelbayev 2017)

  • Utilise a constrained optimization approach to model compression, allowing for the separation of learning and compression processes, thereby enabling the creation of a learning-compression’ algorithm that alternates between learning steps of the uncompressed model and compression steps of the model parameters, irrespective of the compression type or learning task.’ (Carreira-Perpiñán and Idelbayev 2017)

  • Aim to achieve reliable uncertainty from deterministic single-forward pass models, as traditional methods of uncertainty quantification are computationally expensive. (L.-C. Chen et al. 2017)

  • Consider incorporating cross-sample similarities as a form of knowledge transfer in deep metric learning, which can lead to improved performance of student networks. (Yuntao Chen, Wang, and Zhang 2017)

  • Utilise the concept of reshaped tensor decomposition’ to effectively compress neural networks by exploiting inherent invariant structures within them, thereby significantly enhancing your efficiency and applicability across various platforms.’ (Y. Cheng et al. 2017)

  • Consider the impact of non-identical and independent distribution (non-i.i.d.) in your training and testing data sets, and employ techniques like AlignQ to mitigate potential errors caused by these disparities. (Y. Cheng et al. 2017)

  • Consider using a bilevel memory framework with knowledge projection for task-incremental learning, which effectively separates the functions of learning and remembering while ensuring both plasticity and stability. (Y. Cheng et al. 2017)

  • Consider utilising a range of techniques for model compression and acceleration in deep neural networks, including parameter pruning and quantisation, low-rank factorisation, transferred/compact convolutional filters, and knowledge distillation, depending on the specific application and resource limitations. (Y. Cheng et al. 2017)

  • Utilise a novel framework for binary classification based on optimal transport, which incorporates the Lipschitz constraint as a theoretical necessity. This framework proposes to learn 1-Lipschitz networks using a new loss that is an hinge regularised version of the Kantorovich-Rubinstein dual formulation for the Wasserstein distance estimation. This loss function has a direct interpretation in terms of adversarial robustness together with certifiable robustness bounds. (Cisse et al. 2017)

  • Utilize a combination of text-based causal graphs derived from medical literature and observational data from electronic medical records (EMRs) to improve the accuracy and precision of identifying causal relationships among medical conditions., ‘The primary methodological recommendation provided by the paper is to integrate text-based causal graphs from medical literature with observational data from electronic medical records (EMRs) to achieve higher precision in determining causal relationships among medical conditions.’ (D’Amour et al. 2017)

  • Focus on developing hardware accelerators that can efficiently handle variable numerical precision requirements across different layers of deep neural networks, leading to improved performance and energy efficiency. (Delmas et al. 2017)

  • Consider using iterative pruning and re-training to pack multiple tasks into a single deep neural network, thereby avoiding catastrophic forgetting and optimizing for the task at hand. (Fernando et al. 2017)

  • Explicitly model the geometric structure amongst points throughout the hierarchy of feature extraction using a novel convolution-like operation called GeoConv, which helps to preserve the geometric structure in Euclidean space during the feature extraction process. (M. Gao et al. 2017)

  • Consider implementing a reconfigurable scheme for binary neural networks that can dynamically adjust classification accuracy based on specific application requirements, thereby achieving a balance between throughput and accuracy without increasing the area cost of the hardware accelerator. (Ghasemzadeh, Samragh, and Koushanfar 2017)

  • Consider utilizing a style prediction network alongside a style transfer network to enable accurate and efficient predictions of artistic styles for unseen paintings, thereby improving the overall performance of the model. (Ghiasi et al. 2017)

  • Focus on developing computationally efficient deep learning architectures without compromising accuracy, using techniques such as depthwise separable convolutions, parametric rectified linear units, and global average pooling. (T. Ghosh 2017)

  • Consider using Reversible Residual Networks (RevNets) in your studies, as they offer similar classification accuracy to standard ResNets but with significantly lower memory requirements, enabling more efficient training of wider and deeper networks. (Gomez et al. 2017)

  • Employ a hybrid approach combining sparsifying regularizers and uniform width multipliers to optimize deep neural network performance while adhering to resource constraints. (Gordon et al. 2017)

  • Aim to create a continuous relaxation of beam search for end-to-end training of neural sequence models, allowing for improved optimization and better handling of discontinuities in traditional beam search algorithms. (K. Goyal et al. 2017)

  • Consider implementing a quantization scheme that is compatible with training very deep neural networks, where quantizing the network activations in the middle of each batch-normalization module can significantly reduce memory and computational power required, with minimal impact on model accuracy. (B. Graham 2017)

  • Consider using a Sequence-to-Sequence Variational Autoencoder (VAE) for generating vector images, as it provides a robust and flexible framework for handling diverse image classes. (D. Ha and Eck 2017)

  • Use the e-AutoGR framework to improve the explainability of hyperparameter search and performance evaluation strategies in graph representation problems, by using a non-linear hyperparameter decorrelated weighting regression to understand the importance of each hyperparameter in determining model performance. (Hamilton, Ying, and Leskovec 2017b)

  • Utilise a transductive Laplacian-regularised inference for few-shot tasks, which involves minimising a quadratic binary-assignment function comprising both unary and pairwise Laplacian terms, resulting in improved accuracy and efficiency compared to other approaches. (Howard et al. 2017)

  • Utilize channel-wise convolutions to effectively compress deep models, enabling the creation of light-weight CNNs called ChannelNets, which significantly reduce the number of parameters and computational costs without sacrificing accuracy. (Howard et al. 2017)

  • Consider using a recurrent self-analytic STIC trained with VRM and a Gram matrix Regularized MALA (GRMALA) sampler to generate high-quality synthetic images for your analysis. (Howard et al. 2017)

  • Utilise MobileNets, a type of efficient model based on depth-wise separable convolutions, to balance latency and accuracy in mobile and embedded vision applications. (Howard et al. 2017)

  • Aim to develop a fully-aware multi-level attention mechanism that captures the complete information in one text and exploits it in its counterpart layer by layer, resulting in improved accuracy in tasks like machine reading comprehension. (H.-Y. Huang et al. 2017)

  • Consider using a quantization scheme that allows for integer-only arithmetic during inference, which can lead to significant improvements in the latency-versus-accuracy tradeoff for state-of-the-art MobileNet architectures. (Jacob et al. 2017)

  • Optimize neural network queries over video at scale by utilizing inference-optimized model search, which involves searching for and training a sequence of specialized models and difference detectors that preserve the accuracy of the reference network but are specialized to the target video and object, resulting in significant reductions in computational cost. (D. Kang et al. 2017)

  • Use self-normalizing neural networks (SNNs) instead of traditional feed-forward neural networks (FNNs) for better performance, as SNNs automatically converge towards zero mean and unit variance, enabling high-level abstract representations and making learning highly robust. (Klambauer et al. 2017)

  • Utilize submanifold sparse convolutional networks (SS-CNs) for efficient semantic segmentation of 3D point clouds, as demonstrated by your superior performance compared to traditional dense implementations of convolutional networks. (Klokov and Lempitsky 2017)

  • Leverage structured knowledge graphs for visual reasoning when working on multi-label zero-shot learning tasks, as they enable better understanding of the inter-dependencies between seen and unseen class labels. (C.-W. Lee et al. 2017)

  • Focus on developing deep learning architectures that inherently explain your reasoning processes, rather than relying solely on posthoc interpretability analyses. (Chao Li et al. 2017)

  • Carefully differentiate between the roles of 1x1 and kxk convolutions in deep CNNs, and selectively binarize kxk convolutions to create pattern networks that offer significant reductions in model size with minimal impact on performance. (Zhe Li et al. 2017)

  • Consider using Deep Gradient Compression (DGC) to solve the communication bandwidth problem in distributed training by compressing gradients through techniques like momentum correction, local gradient clipping, momentum factor masking, and warmup training, resulting in significant improvements in efficiency without compromising model performance. (Y. Lin et al. 2017)

  • Consider utilizing data-free knowledge distillation for compressing deep neural networks, especially when access to the original training set is unavailable due to privacy, safety, or resource constraints. (Lopes, Fenu, and Starner 2017)

  • Adopt a fine-grained quantization (FGQ) method to effectively convert pre-trained models to a ternary representation, thereby minimizing loss in test accuracy without re-training. (Mellempudi et al. 2017)

  • Consider implementing wide reduced-precision networks (WRPN) in order to balance the trade-off between increasing the number of raw compute operations and reducing the precision of the operands involved in those operations, ultimately leading to improved model accuracy and computational efficiency. (Asit Mishra et al. 2017)

  • Utilise a combination of convolutional networks and knowledge graph embedding methods to effectively answer visual-relational queries in web-extracted knowledge graphs. (Oñoro-Rubio et al. 2017)

  • Incorporate a combination of syntactic and semantic information in the embedding of every word, use a multi-layer memory network for efficient full-orientation matching between the question and context, and leverage a pointer-network based answer boundary prediction layer to accurately identify the location of answers within the passage. (B. Pan et al. 2017)

  • Utilize the Sparse CNN (SCNN) accelerator architecture to enhance the performance and energy efficiency of Convolutional Neural Networks (CNNs) by leveraging the sparsity inherent in the networks weights and activations.’ (Parashar et al. 2017)

  • Utilize relation networks (RNs) as a general purpose neural network architecture for object-relation reasoning, which enables them to effectively learn object relations from scene description data, factorize objects from entangled scene description inputs, and discover implicit relations in one-shot learning tasks. (Raposo et al. 2017)

  • Consider leveraging the inherent robustness of neural networks to tolerate imperfections introduced by lossy weight encoding techniques, such as the Bloomier filter, to achieve significant reductions in memory requirements without sacrificing model accuracy. (Reagen et al. 2017)

  • Adopt a Bayesian point of view in deep learning, incorporate sparsity-inducing priors to prune large parts of the network, and leverage posterior uncertainties to determine the optimal fixed point precision for encoding weights, leading to state-of-the-art compression rates without compromising performance. (Abadi et al. 2016)

  • Utilize “weight sharing” in your architecture search processes to significantly reduce computational costs without sacrificing performance. (B. Baker et al. 2016)

  • Utilize natural-gradient variational inference methods for practical deep learning, leveraging existing techniques such as batch normalization, data augmentation, and distributed training to achieve similar performance in fewer epochs as traditional methods, while still benefiting from the advantages of Bayesian principles. (Bottou, Curtis, and Nocedal 2016)

  • Consider utilizing real-valued non-volume preserving (Real NVP) transformations in your unsupervised learning tasks, as they offer a powerful, stably invertible, and learnable solution for handling high-dimensional data. (Dinh, Sohl-Dickstein, and Bengio 2016)

  • Consider incorporating Bayesian deep learning techniques into your active learning frameworks, particularly when dealing with high-dimensional data such as image datasets, as it allows for better representation of model uncertainty and improved performance overall. (Gutman et al. 2016)

  • Focus on developing techniques to effectively train Quantized Neural Networks (QNNs) with low precision weights and activations, while ensuring minimal loss in prediction accuracy compared to traditional 32-bit counterparts. (Hubara et al. 2016)

  • Utilize a deep learning framework called “Domain Adaptive Hashing” (DAH) to effectively handle unsupervised domain adaptation problems. This involves training a deep neural network to output binary hash codes rather than probability values, which allows for a unique loss function to be developed for target data in the absence of labels and leads to more robust category predictions. (Q.-Y. Jiang and Li 2016)

  • Focus on proving the conjecture for deep linear networks and addressing the open problem for deep nonlinear networks, ultimately leading to a better understanding of the optimization process in deep learning. (Kenji Kawaguchi 2016)

  • Consider combining attention-based and alignment-based methods in your encoder-decoder models for optimal performance in joint intent detection and slot filling tasks. (Bing Liu and Lane 2016)

  • Utilize a combination of channel auto-encoders, domain-specific regularizers, and attention mechanisms to develop efficient and adaptive communication systems capable of handling various channel impairments. (T. J. O’Shea, Karra, and Clancy 2016)

  • Carefully consider the choice of deep learning software tools and hardware platforms, taking into account the specific task requirements and available resources, as different combinations can yield varying levels of performance. (S. Shi et al. 2016)

  • Utilize a conditional variational autoencoder to effectively predict dense trajectories in a scene, thus enabling accurate forecasts of future events. (Walker et al. 2016)

  • Consider using a non-probabilistic variant of the seq2seq model combined with a beam search optimization training procedure to overcome issues of exposure bias, label bias, and loss-evaluation mismatch in sequence-to-sequence learning tasks. (Wiseman and Rush 2016)

  • Focus on developing photonic integrated circuits for ultra-fast artificial neural networks, as they offer a unique combination of interconnectivity and linear operations, making them ideal for high-performance implementations of neural networks. (Yonghui Wu et al. 2016)

  • Consider implementing Trained Ternary Quantization (TTQ) in your deep neural network models to achieve significant reductions in model size without compromising accuracy, thus enabling efficient deployment on mobile devices. (C. Zhu et al. 2016)

  • Consider studying the interactions of multiple flow lines in the context of imaginary geometry, as this provides valuable insights into the properties of these flow lines and your relationship to other random objects with conformal symmetries. (J. Miller and Sheffield 2016)

  • Consider using Long Short-Term Memory-Networks (LSTMNs) for machine reading tasks, as they enable adaptive memory usage during recurrence with neural attention, thereby weakly inducing relations among tokens and improving overall performance compared to traditional methods. (Jianpeng Cheng, Dong, and Lapata 2016)

  • Explore combining low-precision numerics and model compression through knowledge distillation techniques to significantly enhance the performance of low-precision networks. (Song Han, Liu, et al. 2016)

  • Carefully examine the role of implicit regularization in deep learning algorithms, as explicit regularization may not fully explain the generalization error of neural networks. (Szegedy, Vanhoucke, et al. 2016)

  • Consider using the Super-CLEVR virtual benchmark to diagnose the domain robustness of your Visual Question Answering (VQA) models by isolating and studying the impact of four contributing factors: visual complexity, question redundancy, concept distribution, and concept compositionality. (A. Agrawal, Batra, and Parikh 2016)

  • Utilize spectral normalization to effectively stabilize the training process of generative adversarial networks (GANs) by controlling the Lipschitz constant of the discriminator function, thereby improving the overall quality of the generated images. (J. L. Ba, Kiros, and Hinton 2016)

  • Use a combination of deep learning and traditional search methods to improve the accuracy and efficiency of program synthesis, particularly in situations where input-output examples are available. (Balog et al. 2016)

  • Focus on developing simple, carefully designed systems to achieve high levels of accuracy in reading comprehension tasks, as demonstrated by the authors own systems reaching state-of-the-art results of 73.6% and 76.6% on the CNN and Daily Mail datasets.’ (Danqi Chen, Bolton, and Manning 2016)

  • Consider increasing the cardinality’, or the size of the set of transformations, in your deep neural networks as a means to improve classification accuracy while maintaining or reducing complexity.’ (Conneau et al. 2016)

  • Consider implementing Binarized Neural Networks (BNNs) in your deep learning models, as they offer significant improvements in power efficiency due to reduced memory size and accesses, and replacement of most arithmetic operations with bit-wise operations. (Courbariaux et al. 2016)

  • Consider using an Instance Relationship Graph (IRG) for knowledge distillation, as it models three types of knowledge - instance features, instance relationships, and feature space transformation - leading to improved stability, robustness, and performance in comparison to traditional methods. (Courbariaux et al. 2016)

  • Employ dynamic network surgery, involving both pruning and splicing operations, to achieve efficient Deep Neural Networks (DNNs) by balancing network compression and preserving model performance. (Yiwen Guo, Yao, and Chen 2016)

  • Consider using an adaptive version of the straight-through gradient estimator when training binary neural networks, as it can offer superior performance compared to other existing approaches. (E. Jang, Gu, and Poole 2016)

  • Consider combining activation pruning with weight pruning when working with deep neural networks, as this approach can significantly reduce computational costs while preserving model performance. (P. Molchanov et al. 2016)

  • Consider integrating residual connections into your deep convolutional neural networks, as it can lead to significant improvements in training speed and potentially higher recognition performance. (Szegedy, Ioffe, et al. 2016)

  • Consider replacing batch normalization with instance normalization in your deep neural networks for image generation, as doing so can lead to significant improvements in performance. (Ulyanov, Vedaldi, and Lempitsky 2016)

  • Utilise a novel deep kernel learning model combined with a stochastic variational inference procedure to improve classification, multi-task learning, additive covariance structures, and stochastic gradient training in various areas of science. (Wilson et al. 2016)

  • Carefully define attention for convolutional neural networks and utilize this information to enhance the performance of a student network by imitating the attention patterns of a powerful teacher network through attention transfer techniques. (Zagoruyko and Komodakis 2016a)

  • Utilise bias propagation as a pruning technique in deep networks, as it consistently outperforms the traditional approach of merely removing units, irrespective of the architecture and dataset. (Santerne et al. 2016)

  • Utilise dynamic neural computers (DNCs) for efficient and effective handling of complex, quasi-regular structured data, allowing for representation and reasoning about such data while separating large-scale structure from microscopic variability. (Graves et al. 2016)

  • Focus on developing and implementing algorithms for learning displacement operators jointly with the low-rank residual in the low displacement rank (LDR) framework, as it enables the creation of a more general class of LDR matrices that can improve the accuracy of various deep learning applications while reducing the sample complexity of learning. (Anselmi et al. 2016)

  • Focus on identifying linguistic features that are indicative of specific outcomes and decorrelated with confounds, which is crucial for developing transparent and interpretable machine learning NLP models. (Abadi et al. 2016)

  • Utilize TensorFlow, a highly flexible and efficient platform for implementing and deploying large-scale machine learning models, capable of spanning a wide range of hardware platforms and supporting various forms of parallelism. (Abadi et al. 2016)

  • Utilize a Bayesian model that considers the computational structure of neural networks and provides structured sparsity through the injection of noise to neuron outputs while maintaining unregularized weights. (Abadi et al. 2016)

  • Aim to develop end-to-end trainable models that structure your solutions as a library of functions, some of which are represented as source code, and some of which are neural networks, in order to facilitate lifelong learning and efficient knowledge transfer across multiple tasks. (Abadi et al. 2016)

  • Utilise a Generative Adversarial Network (GAN)-based model to transform source-domain images into appearing as if they were drawn from the target domain, thereby improving the performance of unsupervised domain adaptation significantly. (Abadi et al. 2016)

  • Utilise deep neural networks to learn optimization heuristics directly from raw code, rather than relying on hand-crafted features, thereby enabling faster and cheaper heuristic construction. (Abadi et al. 2016)

  • Aim to develop solutions that exploit the structure of deep learning algorithms on two levels: separating and scheduling matrix updates to avoid bursty network traffic, and reducing the size of matrix updates to minimize network load. (Abadi et al. 2016)

  • Consider the impact of real-world distribution shifts on video action recognition models, particularly focusing on the differences between transformer-based and CNN-based models, the benefits of pretraining, and the variability of temporal information importance across datasets. (Abu-El-Haija et al. 2016)

  • Consider utilizing the Visual Interaction Network (VIN) model for predicting future physical states from video data, as it outperforms various baselines and can generate compelling future rollout trajectories. (P. Agrawal et al. 2016)

  • Utilize deep learning algorithms, specifically convolutional neural networks, for premise selection in automated theorem proving, as it outperforms traditional methods and enables efficient handling of large datasets. (Alex A. Alemi et al. 2016)

  • Leverage reinforcement learning to efficiently sample the design space and improve the model compression quality, resulting in significant improvements in accuracy and computational efficiency compared to traditional hand-crafted methods. (Anwar and Sung 2016)

  • Consider using distribution loss to explicitly regulate the activation flow in order to enhance the accuracy of Binarized Neural Networks (BNNs) without compromising your energy advantages. (J. L. Ba, Kiros, and Hinton 2016)

  • Consider using AutoLoss-Zero, a general framework for searching loss functions from scratch for generic tasks, which employs an elementary search space consisting solely of primitive mathematical operators and utilises a variant of the evolutionary algorithm to discover loss functions, improving search efficiency via a loss-rejection protocol and a gradient-equivalence-check strategy. (Bahdanau et al. 2016)

  • Employ a reinforcement learning framework to efficiently search for and prune redundant connections in DenseNet architectures, thereby achieving a better trade-off between accuracy and computational efficiency. (B. Baker et al. 2016)

  • Consider directly compressing range images rather than unprojected point clouds to leverage the lidar scanning pattern, leading to improved compression rates without compromising distortion levels. (Ballé, Laparra, and Simoncelli 2016)

  • Consider using the Re-weighted Adversarial Adaptation Network (RAAN) for unsupervised domain adaptation (UDA) tasks, as it effectively reduces feature distribution divergence and adapts the classifier when domain discrepancies are disparate, achieving state-of-the-art results in extensive evaluations. (Bousmalis et al. 2016)

  • Incorporate relational position encodings into your relational graph attention networks (RGAT) models when studying emotion recognition in conversations (ERC). This allows the model to capture both speaker dependency and sequential information, leading to improved accuracy in recognizing emotions expressed in conversations. (Bradbury et al. 2016)

  • Avoid relying solely on fixed deterministic decompositions of a sequence, especially in areas such as speech recognition, where segmentation should also be informed by the characteristics of the inputs, such as audio signals. Instead, they propose the Latent Sequence Decompositions (LSD) framework, which allows the model to learn a distribution of sequence decompositions and adapt to the specific problem being solved. (W. Chan et al. 2016)

  • Consider utilising a Wide & Deep’ learning framework for recommender systems, which combines the strengths of wide linear models for memorisation and deep neural networks for generalisation, leading to significant improvements in app acquisitions.’ (H.-T. Cheng et al. 2016)

  • Use Hessian-weighted k-means clustering for network quantization to minimize the performance loss due to quantization in neural networks, as it takes into account the varying impact of quantization errors on different network parameters. (Y. Choi, El-Khamy, and Lee 2016)

  • Utilise a high-order residual quantization technique when performing network acceleration tasks, as it offers greater accuracy whilst maintaining the benefits of binary operations. (Courbariaux et al. 2016)

  • Utilise a hierarchical iterative attention model to effectively capture both word level and sentence level information in document-level multi-aspect sentiment classification tasks. (Dhingra et al. 2016)

  • Utilise a “value iteration network” (VIN) - a fully differentiable neural network with a planning module embedded within - to enable your models to learn to plan and thus generalise better to new, unseen domains. (Y. Duan et al. 2016)

  • Consider using conditional instance normalization in style transfer networks to enable the model to learn multiple styles efficiently and effectively, thereby improving the flexibility and applicability of the model. (Dumoulin, Shlens, and Kudlur 2016)

  • Consider using conditional instance normalization in style transfer networks to efficiently model multiple styles simultaneously, allowing for greater flexibility and reduced computational costs. (Dumoulin, Shlens, and Kudlur 2016)

  • Utilise a combination of contrastive learning and adversarial learning techniques to effectively transfer knowledge across different modalities in multi-modal learning systems. (Durugkar, Gemp, and Mahadevan 2016)

  • Consider using a mixture of multiple low-rank factorizations to model a large weight matrix, with the mixture coefficients being computed dynamically depending on the input, in order to improve computation efficiency and maintain (or sometimes outperform) accuracy compared to full-rank counterparts. (D. Ha, Dai, and Le 2016)

  • Consider implementing a dense-sparse-dense (DSD) training approach for deep neural networks to improve optimization performance and reduce overfitting. (Song Han, Pool, et al. 2016)

  • Carefully evaluate the trade-off between network accuracy and hardware metrics like power consumption, design area, and delay when selecting the precision level for neural networks. (Hashemi et al. 2016)

  • Consider the impact of binarization on the loss during the process of binarization itself, rather than just focusing on finding the closest binary approximation of the weights. (L. Hou, Yao, and Kwok 2016)

  • Consider implementing Dense Convolutional Networks (DenseNets) in your studies due to your ability to enhance information flow, mitigate the vanishing-gradient problem, promote feature reuse, and significantly reduce the number of required parameters compared to traditional convolutional networks. (G. Huang et al. 2016)

  • Consider using the Gaussian Context Transformer (GCT) as a highly effective and efficient channel attention block for deep convolutional neural networks, as it enables accurate representation of global contexts through a Gaussian function rather than complex fully-connected layers or linear transformations. (Iandola et al. 2016)

  • Focus on developing algorithms that balance model size, prediction accuracy, and computational efficiency for effective deployment on resource-limited devices. (Daume et al. 2016)

  • Consider implementing local binary convolutional neural networks (LBCNN) as an efficient alternative to standard convolutional neural networks (CNN) for computer vision tasks, as it provides significant parameter savings and computational advantages while maintaining comparable performance. (Juefei-Xu, Boddeti, and Savvides 2016)

  • Consider implementing a two-stage approach for training Bitwise Neural Networks (BNNs): first, conducting traditional network training with a weight compression technique to convert real-valued models into BNNs, followed by performing noisy backpropagation on the resulting BNNs to optimize your performance. (Minje Kim and Smaragdis 2016)

  • Use a combination of exploratory analyses and semi-supervised learning frameworks to identify fraudsters and your strategies in large-scale mobile social networks, taking into account factors such as user demographics, call behavior, and collaboration patterns. (Kipf and Welling 2016a)

  • Consider developing more comprehensive datasets for action quality assessment (AQA) that incorporate multi-person long-form videos with fine-grained annotations, such as the proposed LOGO dataset, to better capture the complexity of real-world scenarios and improve performance in AQA tasks. (Kipf and Welling 2016a)

  • Consider implementing a fully character-level neural machine translation (NMT) model that operates without explicit segmentation, as it allows for improved handling of rare, out-of-vocabulary words and enables efficient multilingual translation. (Jason Lee, Cho, and Hofmann 2016)

  • Focus on developing methods to mitigate the “forgetting catastrophe” in quantization-aware training (QAT) by minimizing the space shift during quantization through proximal quantization space search (ProxQ) and balancing the influence of replay data using a balanced lifelong learning (BaLL) loss function. (Hao Li et al. 2016)

  • Consider using ternary weight networks (TWNs) instead of binary weight networks (BWNs) due to your improved expressive abilities, faster computations, and comparable classification performance on various datasets. (Fengfu Li et al. 2016)

  • Consider using random features instead of relying solely on the kernel trick for efficient learning of Infinite Layer Networks (ILN), as it provides comparable performance without requiring the computation of the kernel. (Livni, Carmon, and Globerson 2016)

  • Consider utilizing the “knowledge distillation” technique, also referred to as “teacher-student” training, in order to enhance the efficiency and effectiveness of your deep learning models. This involves training a compact model under the guidance of a high-performing, complex model, thereby allowing the compact model to benefit from the latters superior capabilities while maintaining its own advantages in terms of size and computational requirements.’ (Liang Lu, Guo, and Renals 2016)

  • Aim to create a comprehensive dataset that enables comparisons between various knowledge sources, including Knowledge Bases (KBs), Information Extraction (IE) pipelines, and raw documents, in order to evaluate the effectiveness of different methods for extracting information and answering questions accurately. (A. Miller et al. 2016)

  • Utilise a unified framework for generalising Convolutional Neural Network (CNN) architectures to non-Euclidean domains like graphs and manifolds, enabling the learning of local, stationary, and compositional task-specific features. (Monti et al. 2016)

  • Utilise self-supervised learning strategies, specifically the Jigsaw puzzle reassembly problem, to effectively teach systems about object composition and spatial arrangements, leading to superior performance in subsequent detection and classification tasks. (Noroozi and Favaro 2016)

  • Utilise a neuro-symbolic program synthesis technique to encode neural search over the space of programs defined using a Domain-Specific Language (DSL). (Parisotto et al. 2016)

  • Utilise unsupervised pretraining to enhance the efficiency of sequence to sequence (seq2seq) models. By initiating the encoder and decoder networks with pretrained weights of two language models and subsequently refining them with labelled data, the authors demonstrate that this strategy substantially boosts the overall performance of seq2seq models. This methodology is particularly advantageous in scenarios where the quantity of supervised training data is limited, thereby reducing the risk of overfitting. (Ramachandran, Liu, and Le 2016)

  • Consider using XNOR-Networks, which involve binarizing both the weights and inputs to convolutional layers, allowing for efficient implementation through XNOR and bitcounting operations, leading to significant speedups and memory savings. (Rastegari et al. 2016)

  • Consider implementing progressive neural networks in your studies, as they enable effective transfer learning without causing catastrophic forgetting, leading to improved performance in various reinforcement learning tasks. (Rusu et al. 2016)

  • Consider using memory-augmented neural networks (MANNs) for one-shot learning tasks, as they have demonstrated superior performance in rapidly assimilating new data and making accurate predictions after only a few samples. (Santoro et al. 2016)

  • Consider extending the teacher-student framework for deep model compression, incorporating a noise-based regularizer when training the student from the teacher, to potentially enhance the performance of the student network. (Sau and Balasubramanian 2016)

  • Utilise an iterative alternating attention mechanism when developing neural attention-based inference models for machine reading comprehension tasks. This mechanism enables the model to explore the query and document in a more fine-grained manner, leading to improved performance compared to traditional methods that collapse the query into a single vector. (Sordoni et al. 2016)

  • Utilise a combination of discriminative modelling, unweighted sharing, and a GAN loss in your adversarial domain adaptation strategies, as demonstrated by the success of the Adversarial Discriminative Domain Adaptation (ADDA) technique. (Taigman, Polyak, and Wolf 2016)

  • Consider utilising graph-structured representations for visual question answering tasks, as this approach significantly improves accuracy compared to traditional CNN/LSTM-based approaches. (Teney, Liu, and Hengel 2016)

  • Create a four-stage process for collecting machine comprehension datasets, specifically focusing on generating exploratory questions requiring reasoning skills, to effectively challenge and improve the capabilities of machine comprehension models. (Trischler et al. 2016)

  • Consider combining match-LSTM and Pointer Net models when developing end-to-end neural networks for machine comprehension tasks, particularly those involving the Stanford Question Answering Dataset (SQuAD). (Shuohang Wang and Jiang 2016)

  • Consider using a multimodal transfer approach, which involves employing a hierarchical deep convolutional neural network that considers both color and luminance channels, and performs stylization hierarchically with multiple losses of increasing scales, to effectively transfer artistic styles onto everyday photographs. (Xin Wang et al. 2016)

  • Explore the concept of cardinality’, defined as the size of the set of transformations, as a crucial dimension alongside the conventional dimensions of depth and width in neural network design. (S. Xie et al. 2016)

  • Consider using a Dynamic Coattention Network (DCN) for question answering tasks, as it enables recovery from initial local maxima corresponding to incorrect answers through an iterative process of focusing on relevant parts of both the question and the document. (C. Xiong, Zhong, and Socher 2016)

  • Focus on developing content-aware neural style transfer algorithms that can effectively distinguish between foreground and background elements in an image, allowing for accurate and realistic style transfers while maintaining the integrity of the original content. (R. Yin 2016)

  • Explore the potential benefits of learning the wavelet filters of scattering networks in 2D signals, rather than relying solely on traditional fixed wavelet filterbank constructions, especially in small-sample classification settings. (Zagoruyko and Komodakis 2016b)

  • Utilise the Gaussian attention model for content-based neural memory access, allowing for greater flexibility in controlling the focus of attention within a neural network, and enabling better handling of semantic distances in latent spaces. (Liwen Zhang, Winn, and Tomioka 2016)

  • Focus on developing methods that leverage true gradient-based learning for binary activated neural networks rather than relying on gradient approximations like the straight through estimator (STE) to achieve higher accuracy and reduce the gap between binary neural networks and your full precision counterparts. (S. Zhou et al. 2016)

  • Consider using a recurrent network to generate model descriptions of neural networks and train this RNN with reinforcement learning to maximize the expected accuracy of the generated architectures on a validation set, leading to improved performance in various domains such as image recognition and language modeling. (Zoph and Le 2016)

  • Utilize a convolutional attentional neural network for extreme summarization tasks, particularly in cases involving source code, due to its ability to effectively capture local time-invariant and long-range topical attention features in a context-dependent manner. (S. Bengio et al. 2015)

  • Consider combining tree-structured Bayesian nonparametric priors with variational autoencoders to enable infinite flexibility of the latent representation space, leading to improved clustering accuracy and generalization capacity. (Bowman, Vilnis, et al. 2015)

  • Consider implementing the hashing trick to achieve significant memory savings while preserving the approximate preservation of inner product operations in your neural network models. (Wenlin Chen et al. 2015)

  • Focus on developing techniques to effectively train Quantized Neural Networks (QNNs) with low precision weights and activations, while still achieving comparable prediction accuracy to your higher precision counterparts. (Zhiyong Cheng et al. 2015)

  • Ensure that your experimental designs promote invariance and disentanglement in deep neural networks by controlling the information in the weights, which can be achieved through implicit or explicit regularization techniques. (Clevert, Unterthiner, and Hochreiter 2015)

  • Consider incorporating unlabelled data in your studies to enhance the stability and generalizability of your models, particularly in cases where labeled data is scarce or expensive. (A. M. Dai and Le 2015)

  • Consider modifying autoencoder neural networks to incorporate autoregressive constraints, allowing for efficient and accurate distribution estimation while maintaining the benefits of a single pass through a regular autoencoder. (M. Germain et al. 2015)

  • Utilise soft targets’, which are essentially smoothened versions of traditional binary classification targets, to enable faster and more accurate learning in deep neural networks.’ (G. Hinton, Vinyals, and Dean 2015)

  • Consider using the SWA-Gaussian (SWAG) method for uncertainty representation and calibration in deep learning, as it provides a simple, scalable, and general-purpose approach that fits a Gaussian using the SWA solution as the first moment and a low rank plus diagonal covariance derived from the SGD iterates, forming an approximate posterior distribution over neural network weights. (Ioffe and Szegedy 2015)

  • Focus on developing unsupervised learning techniques for creating generic, distributed sentence encoders that can effectively represent the meaning and structure of sentences, rather than relying solely on supervised learning methods tailored to specific tasks. (Kiros et al. 2015)

  • Consider employing multi-task learning (MTL) techniques in sequence to sequence models, as demonstrated by the significant improvements observed in translation quality (+1.5 BLEU points) and constituent parsing (93.0 F1 score) when incorporating additional tasks like parsing and image captioning. (M.-T. Luong et al. 2015)

  • Utilise the concept of Hypergradients’, which enables efficient computation of gradients with respect to hyperparameters, thereby facilitating optimization of complex models with numerous hyperparameters.’ (Maclaurin, Duvenaud, and Adams 2015)

  • Utilise the Kronecker-factored Approximate Curvature (K-FAC) method for optimising neural networks, as it offers significant improvements in efficiency and effectiveness over traditional stochastic gradient descent methods. (Martens and Grosse 2015)

  • Focus on improving the model expressiveness and computational efficiency of GMMN through the introduction of adversarial kernel learning techniques, leading to the development of MMD GAN, which significantly outperforms GMMN and is competitive with other GAN works on various benchmark datasets. (F. Yu et al. 2015)

  • Consider incorporating predictive processing into your studies, particularly focusing on interoceptive inference and sensorimotor contingencies, as this approach offers a comprehensive framework for understanding perception, cognition, and action. (Seth 2015)

  • Consider using layer-wise relevance propagation as a general concept for achieving pixel-wise decomposition in non-linear classification architectures, allowing for better interpretability and understanding of complex models. (S. Bach et al. 2015)

  • Consider implementing a two-tiered coarse-to-fine cascade framework for automated computer-aided detection (CADe) in medical imaging, where the first tier generates candidate regions or volumes of interest (ROI or VOI) at high sensitivities but high false-positive (FP) levels, and the second tier employs deep convolutional neural network (ConvNet) classifiers trained on random views of the ROI or V (Roth et al. 2015)

  • Consider using Kronecker Products (KPs) to compress Recurrent Neural Networks (RNNs) for resource-constrained environments, as it allows for significant compression without compromising task accuracy. (Y. Cheng et al. 2015)

  • Avoid pruning by static importance, and instead adopt a dynamic channel pruning strategy that allows the network to learn to prioritize certain convolutional channels and ignore irrelevant ones, thereby accelerating convolution by selectively computing only a subset of channels predicted to be important at runtime. (K. He et al. 2015b)

  • Utilize a novel, gradient-based kernel formulation for noise robustness and an explicit voxel hierarchy structure with compactly supported kernels for scalability when developing a learning-based 3D reconstruction method. (A. X. Chang et al. 2015)

  • Focus on developing specialized hardware for deep learning that utilizes binary weights during forward and backward propagations, while maintaining precision in the stored weights where gradients are accumulated. (Courbariaux, Bengio, and David 2015)

  • Focus on improving the calibration of deep neural networks (DNNs) by incorporating pairwise constraints, which involves providing calibration supervision to all possible binary classification problems derived from the original multiclass problem. (G. Hinton, Vinyals, and Dean 2015)

  • Utilize a novel knowledge distillation method, named CLIPPING, to efficiently transfer the capabilities of a large pre-trained vision-language model to a smaller one, thereby reducing computational costs while maintaining high levels of accuracy. (G. Hinton, Vinyals, and Dean 2015)

  • Consider using data collected from ground vehicles to train a neural network for drone navigation, as it reduces the need for expert drone pilots and increases safety. (Lillicrap et al. 2015)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (Birk et al. 2015)

  • Develop a comprehensive optimization framework that addresses multiple aspects of SNN performance, including reducing SNN operations, enhancing learning quality, quantizing SNN parameters, and selecting appropriate SNN models, in order to achieve memory- and energy-efficient SNNs without sacrificing accuracy. (Diehl and Cook 2015)

  • Consider using a retrain-based quantization method for optimizing the word-length of weights and signals in fixed-point recurrent neural networks, as it demonstrates improved performance when the number of bits is small. (“ICASSP 2016” 2015)

  • Adopt a statistically-grounded pruning criterion for improving the efficiency of deep learning models, as it accounts for parameter estimation uncertainty and leads to enhanced performance and simplified post-pruning re-training. (Z. Tong and Tanaka 2015)

  • Pay attention to the bias term in addition to the gradient when analyzing deep neural networks, as it can significantly impact the accuracy of predictions and provide valuable insights into the models behavior.’ (Russakovsky et al. 2015)

  • Focus on developing techniques for Ensemble Distribution Distillation (EnD²), which involves distilling the distribution of predictions from an ensemble into a single model, thereby enabling the retention of both improved classification performance and information about the diversity of the ensemble, which is essential for accurate uncertainty estimation. (Alipanahi et al. 2015)

  • Utilize the concept of elastic weight consolidation’ (EWC) in your neural network designs to prevent catastrophic forgetting. (Hayashi-Takagi et al. 2015)

  • Focus on developing techniques to effectively train Quantized Neural Networks (QNNs) with low precision weights and activations, while still achieving comparable prediction accuracy to your higher precision counterparts. (Baldassi et al. 2015)

  • Use the “Expected Utility” (EU) metric for evaluating a bidder in online advertising auctions, as it provides a better correlation with A/B test results compared to traditional supervised learning metrics like log likelihood or squared error. (Chapelle 2015)

  • Consider using deep neural networks (DNNs) to build encoding models for understanding the relationship between the hierarchical structure of the ventral visual stream and the complexity of neural population responses. (P. Wang, Malave, and Cipollini 2015)

  • Consider using a combination of multiple LSTMs and a CNN to create a model capable of handling diverse question-answer pairs in a multilingual image question answering system, and evaluate its performance using a Turing Test conducted by human judges. (A. Agrawal et al. 2015)

  • Consider using stacked attention networks (SANs) for image question answering (QA) tasks, as they enable multi-step reasoning and significantly outperform previous state-of-the-art approaches on four image QA data sets. (A. Agrawal et al. 2015)

  • Develop a unified diagram parsing network (UDPnet) that combines object detection and relation matching tasks, along with a dynamic graph generation network (DGGN) that uses dynamic adjacency tensor memory (DATM) to effectively represent and propagate information within a graph structure. (A. Agrawal et al. 2015)

  • Utilise Dynamic Capacity Networks (DCNs) to optimise the efficiency of your deep learning models by dynamically distributing network capacity across an input, thereby reducing computational costs whilst maintaining or even enhancing overall model performance. (Almahairi et al. 2015)

  • Consider implementing a Sparsely-Gated Mixture-of-Experts (MoE) layer in your neural network designs to achieve greater than 1000x improvements in model capacity with minimal impact on computational efficiency. (Amodei et al. 2015)

  • Utilize neural module networks (NMNs) for visual question answering tasks, as they enable the construction of deep networks through the dynamic composition of jointly-trained neural modules based on linguistic structure, leading to improved performance compared to traditional monolithic approaches. (Andreas et al. 2015)

  • Consider utilizing diffusion-convolutional neural networks (DCNNs) for improved predictive performance in working with graph-structured data, due to its flexibility, speed, and accuracy benefits. (Atwood and Towsley 2015)

  • Consider extending Neural Architecture Search (NAS) beyond image classification to dense image prediction, particularly semantic image segmentation, by proposing a network level architecture search space that augments and complements the cell level one, and developing a differentiable, continuous formulation that conducts the two-level hierarchical architecture search efficiently. (Badrinarayanan, Kendall, and Cipolla 2015)

  • Consider incorporating temporal optimization techniques when working with continuous normalizing flows (CNFs) to achieve significant acceleration in training times without sacrificing performance. (Bahdanau, Serdyuk, et al. 2015)

  • Consider developing a quality-of-service-aware neural architecture search (QoS-NAS) framework that enables a single neural network to execute efficiently at various frame rates, offering trade-offs between accuracy and efficiency at minimal latency cost. (E. Bengio et al. 2015)

  • Consider using the Transformer Routing (TRAR) technique to improve the performance of Transformer networks in tasks requiring varying levels of detail, such as Visual Question Answering (VQA) and Referring Expression Comprehension (REC). (E. Bengio et al. 2015)

  • Consider using Variational Network Quantization (VNQ) as a Bayesian network compression method for simultaneously pruning and few-bit quantization of weights in neural networks, resulting in a deterministic feed-forward neural network with heavily quantized weights without the need for additional fine-tuning. (Blundell et al. 2015)

  • Collect a diverse and comprehensive dataset of questions and answers based on a knowledge base, allowing for improved training and evaluation of question answering systems across various domains. (Bordes et al. 2015)

  • Employ a multi-task learning approach on sub-entity granularity to effectively integrate knowledge graphs (KG) with neural machine translation (NMT) models, thereby overcoming issues related to knowledge under-utilization and granularity mismatch. (Bordes et al. 2015)

  • Consider incorporating parameterized hypercomplex multiplication (PHM) layers into your neural network models, as these layers offer greater architectural flexibility and reduced parameter requirements without sacrificing performance. (Bowman, Angeli, et al. 2015)

  • Consider employing a novel cross-modal center loss function alongside other loss functions to effectively eliminate cross-modal discrepancies and enhance the learning of discriminative and modal-invariant features in cross-modal retrieval tasks. (A. X. Chang et al. 2015)

  • Utilise a novel multi-branch attentive feature fusion module in the encoder and an adaptive feature selection module with feature map re-weighting in the decoder to enhance the generalizability of your models. (A. X. Chang et al. 2015)

  • Consider using anchored radial observations (ARO) for learning implicit fields, as it enables accurate and generalizable shape representation by leveraging local shape features and contextual information from multiple viewpoints. (A. X. Chang et al. 2015)

  • Consider utilizing a 3D Generative Adversarial Network (3D-GAN) for generating 3D objects from a probabilistic space. This approach allows for the creation of high-quality 3D objects while enabling the exploration of the 3D object manifold and providing a powerful 3D shape descriptor for 3D object recognition. (A. X. Chang et al. 2015)

  • Leverage pre-trained visual-semantic spaces (VSS) to overcome challenges in scene graph generation (SGG) related to time-consuming ground-truth annotations and limitations in recognizing novel objects outside of training corpora. (Xinlei Chen et al. 2015)

  • Consider utilizing a recurrent neural network (RNN) model to dynamically build a visual representation of a scene while generating captions, allowing for improved results in image caption generation. (Xinlei Chen et al. 2015)

  • Focus on developing function-preserving transformations for neural networks, allowing rapid transfer of knowledge from smaller to larger networks, thereby accelerating the training process and improving overall performance. (Tianqi Chen, Goodfellow, and Shlens 2015)

  • Prioritize locality constraints when scheduling deep learning jobs on multi-tenant GPU clusters, despite potential increased queueing delays, in order to optimize GPU utilization and minimize job runtime. (Tianqi Chen et al. 2015)

  • Adopt a comprehensive approach to optimizing AI pipelines, including leveraging standard APIs, considering the entire pipeline from data preprocessing to deployment, ensuring transparent acceleration, and enabling seamless scalability. (Tianqi Chen et al. 2015)

  • Utilize 8-bit approximation algorithms for parallelizing deep learning tasks, as they effectively compress 32-bit gradients and nonlinear activations, resulting in improved data transfer speeds and maintaining predictive performance on various datasets. (Dettmers 2015)

  • Utilize Winograds minimal filtering algorithms for faster computations in convolutional neural networks, especially when dealing with small filters and small batch sizes.’ (Suyog Gupta et al. 2015)

  • Carefully consider the rounding scheme employed when working with low-precision fixed-point computations in deep neural network training, as stochastic rounding can lead to minimal degradation in classification accuracy compared to standard 32-bit floating-point computations. (Suyog Gupta et al. 2015)

  • Leverage the concept of “cross-modal distillation” to transfer supervision between images from different modalities, allowing for the development of rich representations for unlabelled modalities and serving as a pre-training procedure for new modalities with limited labelled data. (Saurabh Gupta, Hoffman, and Malik 2015)

  • Consider using a channel-wise interaction based binary convolutional neural network learning method (CI-BCNN) for efficient inference, as it effectively addresses the issue of inconsistent signs in binary feature maps resulting from xnor and bitcount operations, thereby preserving information and improving overall performance. (Song Han, Mao, and Dally 2015)

  • Employ a class-aware bilateral distillation method for Few-Shot Class-Incremental Learning (FSCIL) tasks, which involves adaptively drawing knowledge from two complementary teachers - a base model trained on abundant data from base classes and an updated model from the last incremental session - to reduce overfitting risks and prevent catastrophic forgetting. (G. Hinton, Vinyals, and Dean 2015)

  • Prioritize focusing on latency-accuracy tradeoffs instead of FLOPs-accuracy tradeoffs when dealing with few-shot compression scenarios, and that block-level pruning is a superior approach in this context. (G. Hinton, Vinyals, and Dean 2015)

  • Consider extending the contextual encoding layer to 3D point cloud scenarios to better model global contextual information efficiently, while proposing a group contextual encoding method to divide and encode group-divided feature vectors to effectively learn global context in grouped subspaces for 3D point clouds. (Ioffe and Szegedy 2015)

  • Use deep neural networks (DNNs) to extract deep speaker vectors (d-vectors) for semi text-independent speaker verification tasks, as they preserve speaker characteristics and can be effectively combined with conventional i-vector methods. (Lantian Li et al. 2015)

  • Consider implementing a combination of binary (or ternary) connect and quantized back propagation in order to drastically decrease the number of multiplications required in neural network training, potentially leading to improved performance and efficiency. (Zhouhan Lin et al. 2015)

  • Utilize the concept of generalized distillation’, which combines Hinton’s distillation and Vapnik’s privileged information methods, to improve your machine learning models. (Lopez-Paz et al. 2015)

  • Utilize a combination of Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) to effectively generate and comprehend unambiguous object descriptions in images, thereby improving the overall performance of your models. (J. Mao et al. 2015)

  • Develop and utilize advanced visualization tools to gain deeper insight into the complexities of deep neural networks, particularly convolutional neural networks (ConvNets), thereby facilitating improved model designs and overall understanding. (Yosinski et al. 2015)

  • Utilise TensorFlow, a highly adaptable and efficient tool for implementing and deploying large-scale machine learning models, capable of mapping computations onto a wide variety of hardware platforms, thus simplifying the real-world application of machine learning systems. (J. Ba, Mnih, and Kavukcuoglu 2014)

  • Consider replacing traditional Gaussian processes with deep neural networks in Bayesian optimization to achieve better scalability and efficiency, particularly when dealing with high-dimensional problems. (Calandra et al. 2014)

  • Consider implementing deep convolutional networks (DCNs) in fixed point to reduce memory bandwidth, lower power consumption and computation time, and decrease storage requirements for DCNs, especially for real-time processing and deployment on mobile devices or embedded hardware with limited power budgets. (Courbariaux, Bengio, and David 2014)

  • Focus on developing algorithms that can effectively learn from limited amounts of data, particularly in situations where traditional deep learning approaches struggle. (Graves, Wayne, and Danihelka 2014)

  • Aim to build models that are equivariant under transformations of your inputs, such as translations and rotations, in order to improve generalization and reduce sample complexity. (Kanazawa, Sharma, and Jacobs 2014)

  • Consider incorporating multimodal data sources such as images alongside traditional textual inputs in your language models, as it has been shown to improve performance across various tasks including image retrieval from text, text generation from images, and even simple text retrieval. (Kiros, Salakhutdinov, and Zemel 2014)

  • Utilize deep features extracted from various deep learning architectures, as they significantly outperform traditional perceptual metrics in accurately measuring perceptual similarity between images, regardless of the level of supervision employed during training. (Krizhevsky 2014)

  • Carefully consider the trade-off between the ability of a language model to generate novel captions versus its tendency to repeat previously seen captions, as well as the impact of this choice on human perception of the quality of the generated captions. (T.-Y. Lin et al. 2014)

  • Carefully consider the relevance of your chosen dataset and metrics to your intended application domain, and ensure that your experimental setup accurately represents the practical constraints faced in that domain. (Russakovsky et al. 2014)

  • Focus on leveraging the sparsity in bit representations of weights to achieve efficient weight quantization, rather than trying to optimize activations. (Horowitz 2014)

  • Adopt a novel data-driven architecture for predicting human trajectories in future instances, specifically extending Long-Short Term Memory networks (LSTM) for human trajectory prediction, and incorporating a “Social” pooling layer to allow LSTMs of spatially proximal sequences to share your hidden-states with each other. (Bahdanau, Cho, and Bengio 2014)

  • Utilize a novel architecture for neural machine translation that combines a bidirectional RNN as an encoder and a decoder that simulates searching through a source sentence during translation, enabling the model to dynamically attend to different parts of the source sentence and improve overall translation performance. (Bahdanau, Cho, and Bengio 2014)

  • Consider developing multi-layered gradient boosting decision trees (mGBDTs) for improved performance and representation learning abilities, particularly in situations involving discrete or tabular data. (Yoshua Bengio 2014)

  • Utilize knowledge distillation and hint learning to efficiently transfer knowledge from a high-capacity teacher detection network to a compact student network, resulting in improved accuracy and speed for multi-class object detection tasks. (Chatfield et al. 2014)

  • Use multi-level logit distillation, which involves aligning predictions at the instance, batch, and class level, to improve the performance of logit distillation methods in knowledge distillation tasks. (I. J. Goodfellow, Shlens, and Szegedy 2014)

  • Utilise the proposed multi-class N-pair loss’ objective function in deep metric learning tasks, as it enables joint comparison among multiple negative examples, reducing computational burden through an efficient batch construction strategy, and leading to faster convergence and better performance across various visual recognition tasks.’ (Yangqing Jia et al. 2014)

  • Utilise a combination of smoothness-inducing regularisation and Bregman proximal point optimization to manage the complexity of your models and prevent aggressive updating during fine-tuning processes. (Diederik P. Kingma and Ba 2014)

  • Consider implementing Structured Sparsity Learning (SSL) methods in your deep neural networks (DNNs) to enable direct learning of a compressed structure, thereby reducing computation costs and improving classification accuracy. (Simonyan and Zisserman 2014)

  • Consider using highway networks, which enable unimpeded information flow across many layers via adaptive gating units, allowing for the effective training of very deep neural networks through simple gradient descent. (Szegedy et al. 2014)

  • Utilize memory networks, which integrate inference components with a long-term memory component, allowing them to learn how to use these jointly for improved performance in various tasks, particularly in question answering. (Weston, Chopra, and Bordes 2014)

  • Focus on developing algorithms that efficiently approximate complex mathematical functions using simpler, lower-precision representations, allowing for faster and more resource-efficient computation. (Alaghi and Hayes 2014)

  • Create a large-scale distantly supervised challenge dataset for reading comprehension, specifically focusing on complex, compositional questions with syntactic and lexical variability, and requiring cross-sentence reasoning to find answers. (Fader, Zettlemoyer, and Etzioni 2014)

  • Focus on developing novel parametric rectification methods like PReLU, which improve model fitting with minimal additional computation costs and reduce overfitting risks, along with robust initialization methods tailored specifically for rectifier nonlinearities, allowing for successful training of extremely deep rectified models directly from scratch. (F. Agostinelli et al. 2014)

  • Consider using a multilayered Long Short-Term Memory (LSTM) to map input sequences to a fixed-dimensional vector, followed by another deep LSTM to decode the target sequence from the vector, as demonstrated by the authors successful application of this approach to an English to French translation task.’ (Bahdanau, Cho, and Bengio 2014)

  • Consider utilizing an attention-enhanced sequence-to-sequence model for syntactic constituency parsing, as it demonstrates superior performance compared to traditional parsers across various datasets and conditions. (Bahdanau, Cho, and Bengio 2014)

  • Consider the impact of confounding bias caused by the data generation mechanism when developing natural language generation models for courts view generation, and propose a novel Attentional and Counterfactual based Natural Language Generation (AC-NLG) method to mitigate this bias.’ (Bahdanau, Cho, and Bengio 2014)

  • Consider utilizing a bi-directional representation capable of generating both novel descriptions from images and visual representations from descriptions, accomplished through the use of Recurrent Neural Networks (RNNs) and a novel dynamically updated visual representation that serves as a long-term memory of the concepts that have already been mentioned during sentence generation. (Xinlei Chen and Zitnick 2014)

  • Consider using k-means clustering to identify and eliminate redundant spatial patterns within convolutional neural networks (CNNs) in order to improve efficiency and reduce computational requirements without sacrificing accuracy. (Chetlur et al. 2014)

  • Consider using Pointer Networks (Ptr-Nets) for problems requiring variable-length output dictionaries, as demonstrated by your successful application to three complex geometric problems. (Graves, Wayne, and Danihelka 2014)

  • Utilise Deep Neural Decision Forests, a novel approach that combines the strengths of traditional decision trees and deep convolutional networks, allowing for end-to-end training and improved accuracy in machine learning tasks. (Yangqing Jia et al. 2014)

  • Consider using a panoptic lifting scheme based on a neural field representation to generate a unified and multi-view consistent, 3D panoptic representation of a scene, while addressing inconsistencies of 2D instance identifiers across views through a linear assignment with a cost based on the models current predictions and the machine-generated segmentation masks.’ (Diederik P. Kingma and Ba 2014)

  • Consider implementing a novel contrastive visual-textual transformation for sign language recognition (CVT-SLR) to fully leverage the pre-trained knowledge of both the visual and language modalities, leading to improved performance compared to existing single-cue and multi-cue methods. (Diederik P. Kingma and Ba 2014)

  • Consider incorporating a prediction and pattern change detection module into your online MARL algorithms to reduce uncertainty and improve performance in non-stationary environments. (Marinescu et al. 2014)

  • Utilize the concept of knowledge distillation’, which involves training a student network to mimic the output of a larger teacher network, thereby allowing for the creation of smaller, faster-executing models without sacrificing performance.’ (Romero et al. 2014)

  • Focus on developing a deep integration of Convolutional Neural Networks (CNNs) within the MATLAB environment, enabling them to expose CNN building blocks as simple MATLAB commands, thereby facilitating rapid prototyping of new CNN architectures. (Vedaldi and Lenc 2014)

  • Adopt a consensus-based evaluation protocol for image descriptions, which involves comparing the similarity of a candidate sentence to the majority of how most people describe the image, using a triplet annotation modality and the CIDEr metric to capture consensus better than existing choices. (Vedantam, Zitnick, and Parikh 2014)

  • Carefully choose the appropriate neural-embedding model for representing entities and relations in knowledge bases, as different designs can significantly impact the quality of inferences drawn from the data. (Bishan Yang et al. 2014)

  • Distinguish between shallow and deep learners based on the depth of your credit assignment paths, which are chains of potentially learnable, causal links between actions and effects. (Bayer et al. 2013)

  • Consider using a large-scale, structured corpus of over 1 million cooking recipes and 800 thousand food images, called Recipe1M, to train high-capacity models on aligned, multi-modal data, enabling improved performance on tasks such as image-recipe retrieval. (J. Donahue et al. 2013)

  • Utilize the Differentiable Neural Computer (DNC) model for tasks requiring a combination of pattern recognition and symbol manipulation, such as question-answering and memory-based reinforcement learning, due to its ability to manipulate large data structures and learn complex symbolic instructions. (Graves 2013)

  • Utilize the latest available data, particularly from the Large Hadron Collider (LHC), to improve the accuracy of parton distribution functions (PDFs) in particle physics. (Ball et al. 2013)

  • Focus on identifying and studying the effectiveness of various ad-hoc techniques commonly used in the literature for efficient training of binary models, as this will help disambiguate necessary from unnecessary techniques and pave the way for future development of solid theoretical foundations for these. (Yoshua Bengio, Léonard, and Courville 2013)

  • Recognize quantization parameters as directly and jointly learnable parameters during the optimization process, rather than optimizing full-precision weights first and then decomposing them into quantization parameters. (Yoshua Bengio, Léonard, and Courville 2013)

  • Focus on developing methods like A2Q that enable the training of quantized neural networks (QNNs) to use low-precision accumulators during inference without any risk of overflow, thereby increasing the sparsity of the weights and improving the overall trade-off between resource utilization and model accuracy for custom low-precision accelerators. (Yoshua Bengio, Léonard, and Courville 2013)

  • Utilise a learning-based approach rather than a rule-based one when attempting to prune filters in binary neural networks. (Yoshua Bengio, Léonard, and Courville 2013)

  • Focus on developing a novel rate coding SNN-specific attack method called Rate Gradient Approximation Attack (RGA) to improve the effectiveness of adversarial attacks on deep spiking neural networks (SNNs) composed of simple Leaky Integrate-and-Fire (LIF) neurons. (Yoshua Bengio, Léonard, and Courville 2013)

  • Consider using a Binary Graph Convolutional Network (Bi-GCN) to address memory limitations and improve efficiency in graph neural networks (GNNs) without compromising performance. (Yoshua Bengio, Léonard, and Courville 2013)

  • Consider implementing integer-only quantization techniques for Vision Transformers (ViTs) to reduce model complexity and enhance efficient inference on edge devices. (Yoshua Bengio, Léonard, and Courville 2013)

  • Utilise the branch-wise activation-clipping search quantisation (BASQ) methodology to automatically tune the L2 decay weight parameter during the quantisation process of optimised networks, resulting in improved stability and state-of-the-art accuracy. (Yoshua Bengio, Léonard, and Courville 2013)

  • Utilize PeerNets, a novel family of convolutional networks that alternate traditional Euclidean convolutions with graph convolutions, to enhance the robustness of deep learning systems against adversarial attacks. (Bruna et al. 2013)

  • Consider transferring image representations learned with convolutional neural networks (CNNs) on large-scale annotated datasets to other visual recognition tasks with limited training data, as this can lead to significantly improved results for object and action classification, outperforming the current state of the art on Pascal VOC 2007 and 2012 datasets. (J. Donahue et al. 2013)

  • Consider the tradeoff between generality and specificity of features in deep neural networks when conducting transfer learning, as the transferability of features decreases as the distance between the base task and target task increases, but transferring features even from distant tasks can be better than using random features. (J. Donahue et al. 2013)

  • Carefully balance the trade-off between depth, width, filter sizes, and strides in CNN architectures to achieve optimal performance within a constrained time budget. (Eigen et al. 2013)

  • Consider incorporating spatial transformers into your convolutional neural networks to enable active spatial transformation of feature maps, leading to improved performance across various tasks. (I. J. Goodfellow, Bulatov, et al. 2013)

  • Utilise a residual learning framework when dealing with deep neural networks, as it eases the training process and allows for improved accuracy from increased depth. (I. J. Goodfellow, Warde-Farley, Mirza, et al. 2013)

  • Bridge the gap between softmax loss and multi-label scenarios by proposing a multi-label loss function based on relative comparisons among classes, which allows for improved discriminatory power of features and flexibility in application to multi-label settings. (Maji et al. 2013)

  • Focus on developing a scalable matrix factorization approach to learn low-dimensional embeddings for first-order logic formulas, allowing for more accurate and efficient reasoning in artificial intelligence tasks. (Mikolov, Chen, et al. 2013)

  • Consider integrating classification, localization, and detection tasks within a single convolutional neural network (ConvNet) to achieve superior overall performance. (Sermanet et al. 2013)

  • Utilize Theano, a linear algebra compiler that optimizes symbolically-specified mathematical computations, to improve the efficiency of your machine learning models and achieve superior performance compared to alternative libraries like Torch7 and RNNLM. (Bastien et al. 2012)

  • Consider utilizing Deep Neural Networks (DNNs) for acoustic modeling in speech recognition due to your superior performance compared to traditional Gaussian Mixture Models (GMMs) in handling nonlinear manifolds within the data space. (G. Hinton et al. 2012)

  • Consider using large-scale distributed training algorithms like Downpour SGD and Sandblaster L-BFGS to significantly increase the scale and speed of deep network training, ultimately resulting in improved performance on complex tasks such as visual object recognition and speech recognition. (G. E. Dahl et al. 2012)

  • Focus on developing neural networks for end-to-end differentiable proving of queries to knowledge bases by operating on dense vector representations of symbols, allowing for improved performance in handling complex reasoning patterns involving multiple inference steps. (Nickel, Tresp, and Kriegel 2012)

  • Carefully examine the relationship between the choice of label prior model and its potential impact on peaky behavior and convergence behavior during the training process of CTC-based models. (Graves 2012)

  • Utilise random dropout’, wherein a proportion of feature detectors are randomly omitted during training, to prevent complex co-adaptations and thereby reduce overfitting in large feedforward neural networks. (Geoffrey E. Hinton et al. 2012)

  • Consider implementing the hashing trick to achieve significant memory savings while preserving the approximate preservation of inner product operations in your neural network models. (D. C. Cireşan et al. 2011)

  • Utilize neural fields, which are coordinate-based neural networks that parameterize physical properties of scenes or objects across space and time, to effectively solve various visual computing problems and beyond. (Boularias, Kroemer, and Peters 2011)

  • Aim to excel on multiple benchmarks while avoiding task-specific engineering, instead utilizing a single learning system capable of discovering appropriate internal representations across diverse tasks. (Collobert et al. 2011)

  • Consider utilizing the area under the receiver operating characteristic (ROC) curve (Az) as an error measure during the training process of artificial neural networks (ANN)-based classifiers for biomedical data analysis, as it could potentially lead to better performance in terms of Az. (“Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications” 2010)

  • Utilize topological inference and random field theory to analyze complex, smooth, and highly dependent data structures, such as those found in EEG and MEG studies, in order to accurately control for multiple comparisons and improve the reliability of your findings. (Kilner and Friston 2010)

  • Utilise Theano, a math compiler for Python, to improve the speed and efficiency of your machine learning algorithms by up to 44 times, due to its ability to compile mathematical expressions into optimized native machine language. (Bergstra et al. 2010)

  • Carefully manage user expectations regarding the capabilities of automated text recognition systems for historical handwritten documents, taking into account factors like the volume, velocity, variety, and veracity of the data, as well as the limitations of current machine learning techniques. (Bulacu et al. 2009)

  • Utilize a comprehensive taxonomy for categorizing and comparing various feature visualization methods for Convolutional Neural Networks (CNNs), which includes three primary classes: Input Modification, Deconvolutional, and Input Reconstruction methods. (J. Deng et al. 2009)

  • Focus on developing fully-optical neural networks using coherent nanophotonic circuits to achieve significant improvements in computational speed and power efficiency for various learning tasks. (Cardenas et al. 2009)

  • Utilise randomised function fitting algorithms due to your speed and accuracy, despite the lack of theoretical guarantees, as they can approximate various canonical learning algorithms that choose basis functions through costly optimisation processes. (Rahimi and Recht 2008)

  • Consider adopting a modular approach to developing AutoML frameworks, where the generation and evaluation processes are separated into distinct components, enabling greater flexibility, scalability, and ease of comparison between different algorithms. (Floreano, Dürr, and Mattiussi 2008)

  • Consider using codistillation as a distributed training algorithm that utilizes an additional form of communication that is more delay-tolerant, enabling the productive use of more computational resources even beyond the point where adding more workers provides no additional speedup for SGD. (“Proceedings of the 23rd International Conference on Machine Learning - ICML ’06” 2006)

  • Focus on accurately defining the network knowledge in order to optimize the performance of the distilled network. (Buciluǎ, Caruana, and Niculescu-Mizil 2006)

  • Adopt a hierarchical Bayesian inference framework for studying the visual cortex, which allows for the integration of top-down contextual priors and bottom-up observations to perform concurrent probabilistic inference along the visual hierarchy. (T. S. Lee and Mumford 2003)

  • Focus on developing a novel motion descriptor that disentangles the standard pose representation by removing subject-specific features, which will improve the generalizability of your models when dealing with soft-tissue dynamics. (B. Allen, Curless, and Popović 2003)

  • Analyze the behavior of deep neural networks (DNNs) using an information theoretic approach, specifically focusing on the mutual information between layers and the input variable, and the desired label, during the training dynamics. (Paninski 2003)

  • Utilize the collective wisdom within the neural networks published in online code repositories to create better reusable neural modules, thereby reducing the complexity and cost of subsequent neural architecture creation policies. (X. Yan and Han 2003)

  • Consider the potential differences between various artificial grammar systems, as well as the importance of controlling for factors such as vocabulary size and interference between languages, in order to better understand the neural basis of artificial grammar learning. (Skosnik et al. 2002)

  • Consider utilizing automated machine learning (AutoML) techniques throughout the machine learning pipeline, particularly focusing on neural architecture search (NAS) for optimal model generation, while addressing open problems and exploring future directions in the field. (Stanley and Miikkulainen 2002)

  • Utilize a neural network model instead of traditional linear or logistic regression models when studying international conflicts due to the complexity and rarity of the phenomenon, allowing for more accurate predictions and identification of significant factors. (N. Beck, King, and Zeng 2000)

  • Use a hierarchical model with a MAX-like operation to account for complex visual tasks such as object recognition, as it is consistent with physiological data from inferotemporal cortex and makes testable predictions. (Riesenhuber and Poggio 1999)

  • Explore the potential benefits of utilizing extended context in attention-based neural machine translation, particularly in improving textual coherence and translation quality. (Hochreiter and Schmidhuber 1997)

  • Aim to develop equivariant scene representations for neural rendering, which means ensuring that the learned representation transforms like a real 3D scene, thus improving the accuracy and efficiency of the rendering process. (Curless and Levoy 1996)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (G. Lowe 1995)

  • Consider using EnSyth, a deep learning ensemble approach, to improve the predictability of compact neural network models by generating a diverse set of compressed models using different hyperparameters for a pruning method, synthesizing your outputs via ensemble learning, and exploring the best performing combinations of models using backward elimination. (Girosi, Jones, and Poggio 1995)

  • Consider a broader spectrum of representational schemes when studying intelligent behavior, moving beyond the binary of explicit versus implicit representation and recognizing a rich continuum of degrees and types of representationality. (Andy Clark and Toribio 1994)

  • Utilise a three-step process for refining existing knowledge using neural networks: first, insert knowledge into a neural network, second, refine the network through standard neural learning techniques, and third, extract refined knowledge from the network. (Towell and Shavlik 1993)

  • Utilize Bayesian methods for adaptive models, as they effectively embody Occams Razor, allowing for the automatic identification of over-complex and under-regularized models as less probable, despite your potential to fit the data better.’ (MacKay 1992a)

  • Utilize a Bayesian framework for backpropagation networks, which enables them to make objective decisions regarding network architecture, weight decay rates, and model selection while incorporating Occams Razor to prevent overfitting.’ (MacKay 1992b)

  • Focus on developing frameworks for quantifying the robustness of neural networks to parameter quantization, enabling safer deployment of neural networks on edge devices. (Rumelhart, Hinton, and Williams 1986)

  • Carefully consider the rounding scheme employed when working with low-precision fixed-point computations, as it plays a crucial role in determining the networks behavior during training.’ (Kung 1982)

  • Utilise entropy penalised reparameterisation for scalable model compression, allowing for improved classification accuracy and model compressibility simultaneously. (Rissanen and Langdon 1981)

  • Consider using functional correctness as a metric for evaluating generative models for code, as opposed to traditional match-based metrics, as it accounts for the vast space of functionally equivalent programs and aligns with how humans judge code quality. (Manna and Waldinger 1971)

  • Consider using locally constant networks, which are based on ReLU networks, to effectively and efficiently represent and train oblique decision trees, leading to improved performance in various applications. (Vapnik and Chervonenkis 1971)

  • Utilise the Gauss-Newton approximation to the Hessian matrix within the Levenberg-Marquardt algorithm for efficient implementation of Bayesian regularisation in the training of feedforward neural networks. (Foresee and Hagan, n.d.)

  • Focus on developing self-organizing neural networks capable of recognizing patterns based on geometric similarity while being unaffected by shifts in position or minor changes in shape or size. (NA?)

  • Focus on developing a precise and quantitative formulation of the laws governing the dynamics of individual neurons and your interactions in large neuronal assemblies, using a simplified model of the real system based on abstraction and trial-and-error. (NA?)

  • Carefully review your work for potential errors and inconsistencies, such as incorrect formulas or misplaced figures, and ensure they accurately represent your findings. (NA?)

  • Modify the Hebbian model of classical conditioning by incorporating changes in pre- and postsynaptic levels of activity, sequentially correlating these changes, and making the change in synaptic efficacy proportional to its current efficacy, leading to a more accurate prediction of various animal learning phenomena. (NA?)

  • Ensure your studies are designed to capture the essential elements of the phenomenon being studied, taking into account factors such as sample size, measurement validity, and statistical power. (NA?)

  • Correct the proof of Lemma 1 in Cybenkos original paper by replacing instances of \(L^\infty(\mathbb{R})\) with \(L^\infty(J)\) for a compact interval \(J\) containing \(\{y^T x | x \in I_n\}\), where \(y\) is fixed, and noting that the reduction of multidimensional density to one-dimensional density was previously achieved by Dahmen and Micchelli in your work on ridge regression.’ (NA?)

  • Utilise a three-step process for refining existing knowledge using neural networks: first, insert knowledge into a neural network; second, refine the network using standard neural learning techniques; and finally, extract refined knowledge from the network. (NA?)

  • Carefully differentiate between type-1 and type-2 problems, as type-2 problems require the exploitation of indirect justifications involving the derivation of a recoding of the training examples and the derivation of probability statistics within the recoded data, while type-1 problems can be solved through the exploitation of observable statistical effects in the input data. (NA?)

  • Focus on developing simulations that explore the co-evolution of language production and comprehension abilities in populations of neural networks, emphasizing the importance of understanding the selective pressures driving the evolution of these abilities. (NA?)

  • Carefully curate your training datasets, removing homologous sequences and checking against primary sources, to avoid bias and improve the performance of machine learning algorithms. (NA?)

  • Utilize soft computing methodologies, such as fuzzy sets, neural networks, genetic algorithms, and rough sets, in conjunction with traditional techniques, to effectively tackle the numerous challenges associated with data mining, including massive data sets, high dimensionality, user interaction, overfitting, understandability of patterns, nonstandard and incomplete data, mixed media data, and management of changing data and knowledge. (NA?)

  • Focus on developing comprehensive models that incorporate both the primacy gradient and response suppression mechanisms, allowing them to better understand and predict various aspects of serial recall. (NA?)

  • Carefully consider the choice of input variables when developing artificial neural networks (ANNs), as it affects model complexity, learning difficulty, and performance, and employ appropriate variable selection methods to optimize the ANN model. (NA?)

  • Focus on improving the performance of your predictive models through the incorporation of additional relevant features, rigorous error correction of datasets, and regular updates to algorithm components. (NA?)

  • Allow evolution to complexify, i.e., to incrementally elaborate on solutions through adding new structure, in order to discover and improve complex solutions. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Avoid imposing arbitrary classification boundaries on real-valued variables like solvent accessibility, and instead opt for continuous approximation methods like nonlinear regression using neural networks. (NA?)

  • Focus on testing the “Bayesian coding hypothesis” through experimental approaches, specifically examining whether and how neurons code information about sensory uncertainty. (NA?)

  • Carefully consider the choice of regularization techniques and early stopping strategies when working with perceptrons, multi-layer perceptrons, and support vector machines, as they significantly influence the margin and generalization capabilities of these models. (NA?)

  • Consider utilizing a combination of statistical phrase extraction and neural network-based self-organizing map (SOM) categorization to effectively generate hierarchical knowledge maps from large volumes of textual data, such as online news articles, thereby enabling users to efficiently browse and discover relevant information. (NA?)

  • Consider using a cooperative coevolutionary approach for designing neural network ensembles, which involves simultaneously evolving both the individual networks and your combinations, while evaluating each networks performance using a multi-objective method that considers not just its performance in the given problem, but also its cooperation with the rest of the networks.’ (NA?)

  • Consider using stabilized finite element methods when dealing with certain types of differential equations, particularly those involving convection operators, as these methods can lead to more accurate and reliable solutions. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Prioritize making frequent but smaller updates to your model parameters during the training phase, as opposed to infrequent but larger updates, in order to achieve optimal results in machine translation tasks. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Utilise the LambdaRank algorithm to improve the efficiency and effectiveness of your ranking models, especially when dealing with nonsmooth cost functions. (NA?)

  • Consider utilizing Echo State Networks (ESNs) instead of Simple Recurrent Networks (SRNs) when working on natural language tasks, as ESNs demonstrate comparable performance without requiring extensive training of internal representations. (NA?)

  • Consider evaluating deep learning algorithms on more complex problems with many factors of variation, rather than just simpler ones like digit recognition, to better understand your capabilities and limitations. (NA?)

  • Utilize the free-energy principle, which involves minimizing the difference between expected and actual sensory input, to better understand the organization and response patterns of complex systems like the brain. (NA?)

  • Consider utilizing Support Vector Machines (SVMs) for neuroimaging-based diagnosis due to its potential for achieving higher accuracy rates than human radiologists, particularly in areas where trained experts are scarce. (NA?)

  • Employ a Bayesian approach to compressive sensing, which enables them to estimate both the underlying signal and its error bars, determine when enough measurements have been taken, optimize compressive sensing measurements adaptively, and account for additive noise in the measurements. (NA?)

  • Utilise advanced machine learning techniques, particularly convolutional neural networks, to effectively analyse complex, high-dimensional spatiotemporal patterns of EEG synchronisation for improved seizure prediction accuracy. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Utilise second-order Markov logic for deep transfer learning tasks, enabling the discovery of structural regularities in the source domain through Markov logic formulas with predicate variables, which can then be instantiated with predicates from the target domain. (NA?)

  • Utilise a Deep Boltzmann Machine (DBM) for multimodal learning, as it enables the extraction of a unified representation that fuses multiple and diverse input modalities together, which is beneficial for classification and information retrieval tasks. (NA?)

  • Carefully analyze the impact of diversity within online ensemble learning systems during periods of concept drift, as it can significantly affect performance and adaptation capabilities. (NA?)

  • Consider utilising advanced computer vision and machine learning algorithms to develop automated systems capable of accurately recognising and analysing various aspects of mouse behaviour within your natural environment, thereby providing valuable insights into your phenotypes and facilitating large-scale studies. (NA?)

  • Utilise Sensitivity Analysis (SA) methods to enhance the interpretability of black box’ data mining models like Neural Networks, Support Vector Machines, and Random Forests.’ (NA?)

  • Utilize the free-energy formulation of active inference to understand the mirror-neuron system, as it allows for the simulation of neuronal processes involved in action-observation and the generation of motor behavior. (NA?)

  • Consider implementing a reservoir computer in which the usual structure of multiple connected nodes is replaced by a dynamical system comprising a nonlinear node subjected to delayed feedback, as this approach provides excellent performance on benchmark tasks while requiring fewer components to build. (NA?)

  • Ensure the full column rank of the hidden layer output matrix H in your neural network model to improve the learning rate, testing accuracy, prediction accuracy, and overall robustness of the network. (NA?)

  • Consider integrating neuron division and budding mechanisms into spiking neural P systems to improve your efficiency and enable them to solve computationally difficult problems in polynomial time. (NA?)

  • Consider the importance of developing a comprehensive and flexible architecture for the Internet of Things (IoT) that addresses issues such as scalability, interoperability, reliability, Quality of Service (QoS), and security, while also considering the potential impact of IoT on various industries and aspects of daily life. (NA?)

  • Consider utilizing a multi-stage machine learning approach with increasingly refined levels of resolution for improved protein contact map prediction. (NA?)

  • Adopt the framework of active inference, wherein the motor system sends descending proprioceptive predictions rather than motor commands, allowing for a more nuanced understanding of the complex interactions between the motor and sensory systems. (NA?)

  • Consider implementing a “grow when required” (GWR) network for unsupervised learning tasks, which dynamically adjusts its structure based on the input data, leading to improved accuracy and efficiency in mapping high-dimensional input spaces to lower-dimensional representations. (NA?)

  • Consider employing a combination of multiple forecasting models, including numerical weather prediction, ensemble forecasting, upscaling and downscaling processes, statistical and machine learning approaches, to enhance the accuracy and robustness of wind power forecasting. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Utilize hierarchical predictive coding strategies in your studies, which involve the use of top-down probabilistic generative models to predict the flow of sensory data, thereby allowing them to make accurate inferences about the signal source (or the world) based on the varying input signal alone. (NA?)

  • Utilize hierarchical predictive processing models to understand how the brain uses top-down generative models to make accurate predictions about the environment, thereby reducing prediction error and improving perception and action. (NA?)

  • Focus on analyzing the dynamics of neural microcircuits from the perspective of a readout neuron, which can learn to extract salient information from the high-dimensional transient states of the circuit and transform transient circuit states into stable readouts, allowing for invariant readout despite the lack of repeated states. (NA?)

  • Focus on developing deep learning methods for representation learning, which aim to create more abstract and useful representations of data by composing multiple nonlinear transformations, thereby enabling better understanding of the underlying explanatory factors and improving the performance of machine learning algorithms. (NA?)

  • Consider using a semantic matching energy function to effectively embed multi-relational data into a flexible continuous vector space, allowing for accurate predictions and efficient manipulation of large-scale structured data across diverse applications. (NA?)

  • Focus on developing algebraic structures for combining previously acquired knowledge through trainable modules, rather than attempting to bridge the gap between machine learning systems and advanced inference mechanisms. (NA?)

  • Consider combining rank-order learning and dynamic synapses in evolving spiking neural networks (eSNN) to efficiently recognize spatio- and spectro-temporal data (SSTD) in an online mode. (NA?)

  • Utilise Support Vector Machine (SVM) classifiers along with mobile EEG sensors to distinguish between attentive and inattentive states in students during learning processes. (NA?)

  • Focus on designing specialized, efficient hardware for specific machine-learning algorithms, rather than attempting to create general-purpose solutions. (NA?)

  • Consider utilizing deep learning techniques, specifically deep neural networks (DNNs), for improved performance in signal and information processing tasks, particularly when dealing with complex natural signals like human speech, natural sounds, languages, images, and visual scenes. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Use a hybrid intelligent algorithm (HIA) approach combining extreme learning machine (ELM) and particle swarm optimization (PSO) to directly formulate optimal prediction intervals of wind power generation, thereby improving accuracy and reliability while reducing the need for prior knowledge, statistical inference, or distribution assumptions about forecasting errors. (NA?)

  • Use the Extreme Learning Machine (ELM) combined with the pairs bootstrap method for probabilistic forecasting of wind power generation, as it effectively accounts for the uncertainties in the forecasting results and provides a high potential for practical applications in power systems. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Consider implementing a passive photonic silicon reservoir for ultrafast, low-power optical information processing, as it can effectively handle both digital and analogue tasks while consuming minimal energy. (NA?)

  • Carefully control for experimental limitations and computational considerations when comparing the representational performance of deep neural networks (DNNs) to that of the primate visual system, using methods like kernel analysis to ensure a fair comparison. (NA?)

  • Carefully select appropriate sensors and electrodes for measuring hand kinematics, dynamics, and muscular activity, ensuring proper placement and synchronization of data streams, and utilizing advanced signal processing techniques such as filtering and relabeling to enhance the quality and reliability of collected data. (NA?)

  • Utilize the proposed structure2vec’ method for efficient and accurate handling of structured data, particularly in scenarios involving millions of data points, due to its ability to effectively combine graphical models, embedding techniques, and discriminative training.’ (NA?)

  • Consider adopting the Extreme Learning Machine (ELM) algorithm instead of the Artificial Neural Network (ANN) algorithm for predicting the Effective Drought Index (EDI) in Eastern Australia because it demonstrates superior performance in terms of prediction accuracy, learning speed, and training speed. (NA?)

  • Utilise the eigenbrain method when conducting studies involving Alzheimers disease (AD) subject prediction and discriminate brain-region detection in MRI scanning due to its demonstrated efficacy.’ (NA?)

  • Consider applying deep learning algorithms to address specific problems in big data analytics, such as learning from massive volumes of data, semantic indexing, discriminative tasks, and data tagging, while also focusing on improving specific areas of deep learning to accommodate challenges associated with big data analytics, such as learning from streaming data, dealing with high dimensionality of data, scalability of models, and distributed and parallel computing. (NA?)

  • Consider the depth of credit assignment paths (CAPs) when evaluating the effectiveness of deep learning algorithms in neural networks, as deeper CAPs indicate greater potential for improved performance in future episodes. (NA?)

  • Combine multiple data sources, including MRI, age, and cognitive measures, when developing models to predict the likelihood of MCI patients converting to Alzheimers disease.’ (NA?)

  • Consider utilizing integrated photonic tensor cores for parallel convolution processing, as they offer the advantage of operating at Tera-Multiply-Accumulate per second (TMAC/s) speeds, reducing computation to measuring the optical transmission of reconfigurable and non-resonant passive components, and operating at a bandwidth exceeding 14 GHz, limited only by the speed of the modulators and photodetectors. (NA?)

  • Consider utilising a combination of data-augmented classification along with radiomics hypothesis to improve the accuracy of prostate cancer diagnoses, thus potentially reducing the chances of under- or overdiagnosis. (NA?)

  • Utilize a large dataset of vector magnetograms, combined with a nonlinear classification algorithm like Support Vector Machines (SVM), to achieve improved predictive accuracy when attempting to forecast solar flares. (NA?)

  • Focus on optimizing objective functions, learning rules, and architectures in order to better understand and model complex neural systems. (NA?)

  • Consider implementing a maximum entropy based confidence penalty and label smoothing as regularizers for large, deep neural networks, as these techniques have been shown to improve state-of-the-art models across various benchmarks without requiring modification of existing hyperparameters. (NA?)

  • Utilise a deep learning approach for network intrusion detection in software defined networking (SDN) environments, specifically through building a Deep Neural Network (DNN) model and training it with the NSL-KDD Dataset. (NA?)

  • Strive to create machines that learn and think like humans by focusing on three main elements: building causal models of the world, grounding learning in intuitive theories of physics and psychology, and leveraging compositionality and learning-to-learn to rapidly acquire and generalize knowledge to new tasks and situations. (NA?)

  • Consider combining convolutional neural networks (CNNs) with multiple instance learning (MIL) when working with microscopy images, enabling accurate classification and segmentation without needing explicit segmentation steps or single cell level labelling. (NA?)

  • Consider using resistive processing units (RPUs) to accelerate deep neural network (DNN) training by orders of magnitude while reducing power consumption, enabling faster and more efficient large-scale analysis and classification tasks. (NA?)

  • Consider fine-tuning pre-trained deep convolutional neural networks (CNNs) instead of training them from scratch for medical image analysis, as it offers better performance, increased robustness to training set sizes, and a flexible layer-wise fine-tuning scheme tailored to the amount of available data. (NA?)

  • Consider utilizing smartphone sensors and machine learning algorithms to develop context-aware digital therapies for people with depression, offering in-situ support while maintaining privacy and minimizing intrusion. (NA?)

  • Consider utilizing transfer learning techniques to leverage existing large datasets from one domain (such as mammography) to improve the accuracy of deep convolutional neural networks in another related domain (like digital breast tomosynthesis), thus reducing the need for extensive data collection efforts. (NA?)

  • Consider using the eigendecomposition of the Laplace operator as a unifying mathematical framework to understand and predict the collective dynamics of human cortical activity at the macroscopic scale. (NA?)

  • Consider using a GPU-specialized parameter server, such as GeePS, to overcome the limitations of traditional CPU-based parameter servers in supporting scalable deep learning across distributed GPUs. (NA?)

  • Utilise a 10-fold cross validation technique when testing your models, ensuring that all algorithms share the same sample partition settings on each fold for fair comparisons. This approach allows for accurate evaluation of the performance of different algorithms across multiple iterations, providing robust evidence for any conclusions drawn. (NA?)

  • Carefully select and combine various texture descriptors and classifiers to improve the accuracy of multiclass tissue classification tasks in histopathological images. (NA?)

  • Consider utilizing convolutional neural networks (CNNs) for efficient and accurate cancer detection in histopathology, particularly in scenarios where traditional methods may be labor intensive or prone to human error. (NA?)

  • Consider utilizing unsupervised deep feature learning to create a more comprehensive and accurate representation of Electronic Health Records (EHRs) for predictive clinical modelling purposes. (NA?)

  • Utilise machine learning algorithms, specifically reservoir computing, to estimate the Lyapunov exponents of a chaotic process from limited time series data. (NA?)

  • Consider employing machine learning (ML) accelerated ab initio molecular dynamics (AIMD) simulations to improve the efficiency and accuracy of simulating vibrational spectra in complex molecular systems. (NA?)

  • Focus on understanding the mathematical foundations of deep learning algorithms, explore various applications of recurrent neural networks, and consider using advanced techniques like Monte Carlo methods and partition functions for better feature representation and optimization. (NA?)

  • Develop a generative model based on a deep recurrent architecture that combines recent advances in computer vision and machine translation to accurately describe the content of an image using natural language. (NA?)

  • Consider employing deep neural networks (DNNs) for modeling bioactivity data, particularly when using the rectified linear units (ReLU) activation function, having at least two or three hidden layers, optimizing the number of neurons per hidden layer on a case-by-case basis, and applying dropout regularization to both input and hidden layers. (NA?)

  • Focus on selecting graph neural networks with greater depth and width when dealing with complex graph classification tasks, as restricting these parameters can lead to significant loss of expressive power and make certain decision problems impossible to solve. (NA?)

  • Integrate rematerialization and paging techniques to effectively reduce memory consumption of large, state-of-the-art ML models, allowing for energy-efficient training on memory-scarce battery-operated edge devices. (NA?)

  • Consider using multiple processors, especially GPUs, to achieve higher efficiency and speed when working with large datasets and complex models in machine learning applications. (NA?)

  • Carefully consider the timing, location, and method of sparsifying neural networks to achieve optimal computational efficiency and model accuracy. (NA?)

  • Consider leveraging deep neural networks (DNNs) to automatically learn effective patterns from categorical feature interactions in user response prediction, particularly in areas like web search, personalized recommendation, and online advertising. (NA?)

  • Consider incorporating network morphism into genetic algorithms for optimizing neural architecture search in medical image classification tasks, as it can help reduce running time and improve overall model performance. (NA?)

  • Leverage the inherent structure and simplicity found in real-world datasets, such as symmetry, locality, compositionality, and polynomial log-probability, to create highly efficient and effective deep learning models. (NA?)

  • Utilize Convolutional Neural Networks (CNNs) in your studies, as they offer robustness to misalignment issues and can effectively handle the PoI selection problem and misalignment issue simultaneously. (NA?)

  • Focus on developing deep learning algorithms that enable spatially and chemically resolved insights into quantum-mechanical properties of molecular systems beyond those trivially contained in the training dataset, while maintaining interpretability, size-extensiveness, efficiency, and uniform accuracy across compositional and configurational chemical spaces. (NA?)

  • Consider implementing a flexible, 3D stacking, artificial chemical synapse network (3D-ASN) using selector-device-free electronic synapses (e-synapses) to effectively mimic correlated learning and exhibit a trainable memory function with a strong tolerance to input faults. (NA?)

  • Use a write-verify programming scheme for your neural networks to achieve faster convergence and improved accuracy in tasks like face classification. (NA?)

  • Leverage the wealth of knowledge available in neuroscience to inform and validate the development of artificial intelligence algorithms and architectures, thereby improving the likelihood of creating truly intelligent machines. (NA?)

  • Utilize Convolutional Neural Networks (CNNs) for the classification of hematoxylin and eosin stained breast biopsy images, as this method retrieves information at various scales, enabling accurate identification of normal tissue, benign lesions, in situ carcinoma, and invasive carcinoma. (NA?)

  • Utilize Quantum Loop Topography (QLT) as a means of converting complex quantum information into a format suitable for analysis by a neural network. (NA?)

  • Consider implementing the NICE (Noise Injection and Clamping Estimation) method for neural network quantization, which involves noise injection during training to mimic quantization noise and statistics-based initialization of parameter and activation clamping for faster model convergence. (NA?)

  • Consider using a dynamic termination state in your neural network architectures, allowing the system to adaptively determine when to stop reading and start producing an answer based on the complexity of the input data. (NA?)

  • Carefully consider the choice of word representation (word-based vs. character-based), encoder depth, target language, and encoder vs. decoder representations when evaluating the quality of neural machine translation (NMT) models for learning morphology. (NA?)

  • Utilize the Tensor Algebra Compiler (TACO) to automatically generate kernels for any compound tensor algebra operation on dense and sparse tensors, improving performance and saving memory compared to manual implementation. (NA?)

  • Utilize leave-one-out cross-validations to ensure unbiased training and testing, while also considering the impact of confidence scores on precision and recall when evaluating the performance of various de novo sequencing tools. (NA?)

  • Consider incorporating advanced computational brain network modeling techniques, such as the Hopf model, to better understand the complex spatio-temporal dynamics of brain function and improve the accuracy of your findings. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Explore the use of deep learning techniques, specifically deep convolutional neural networks (DCNNs), for blind image quality assessment (BIQA), as they offer superior performance when compared to traditional methods. (NA?)

  • Carefully consider the trade-off between accuracy and computational efficiency when developing deep neural networks, particularly in the context of reducing precision and utilizing quantization techniques. (NA?)

  • Consider utilizing deep neural networks for predicting fluorescent labels from transmitted-light images, as demonstrated by the successful application of this method in distinguishing various cell types and structures. (NA?)

  • Aim to design your reservoir systems to yield a one-to-one synchronization function, which guarantees the existence of a function that maps the reservoir state to the measurement vector, allowing for accurate short-term forecasts and long-term climate replication. (NA?)

  • Combine knowledge-based models with machine learning techniques to create a hybrid forecasting scheme for improved accuracy and wider applicability in predicting chaotic processes. (NA?)

  • Consider developing and implementing automated methods for interpreting echocardiograms, which could potentially improve access to cardiac evaluations in primary care settings and rural areas while reducing costs and improving efficiency. (NA?)

  • Consider integrating deep learning approaches with machine learning techniques for improved accuracy in short-term load forecasting (STLF) tasks, as demonstrated by the superior performance of the proposed deep neural network algorithm compared to five commonly used artificial intelligence algorithms. (NA?)

  • Utilise the Conditional Variational Autoencoder (CVAE) model for molecular design tasks, as it allows for simultaneous control of multiple target properties, thus enabling efficient molecular design. (NA?)

  • Carefully evaluate the suitability of deep learning methods for your specific biomedical problem, considering factors like data availability, quality, and relevance, as well as the need for interpretable models and efficient representation of underlying data structures. (NA?)

  • Consider implementing a multi-memristive synaptic architecture with an efficient global counter-based arbitration scheme to effectively manage the conductance modulation of memristive devices in artificial neural networks, thereby enhancing the accuracy and scalability of neuromorphic computing systems. (NA?)

  • Consider implementing in-situ learning in multi-layer memristor neural networks for efficient and self-adaptive processing, particularly when dealing with complex datasets like MNIST handwritten digits. (NA?)

  • Consider employing unsupervised machine learning methods when working with large, diverse datasets to avoid subjectivity in feature selection and potentially achieve improved classification accuracy. (NA?)

  • Combine network analysis with behavioral properties to effectively detect fraudulent users in online platforms. (NA?)

  • Consider the potential for heterogeneity within the frontoparietal control network (FPCN) and explore its relationship with the default mode network (DMN) and dorsal attention network (DAN) through hierarchical clustering and machine learning classification analyses of within-FPCN functional connectivity patterns. (NA?)

  • Carefully consider the composition and representativeness of your training sets when developing automated diagnostic systems for pigmented skin lesions, ensuring adequate coverage of various disease classes and minimizing bias towards certain conditions. (NA?)

  • Adopt a deep learning method for microstructural classification in steel, specifically through the use of pixel-wise segmentation via Fully Convolutional Neural Networks (FCNN) combined with a max-voting scheme, as this approach significantly improves classification accuracy compared to existing methods. (NA?)

  • Utilize the proposed all-optical diffractive deep neural network (D^2NN) architecture for performing machine learning tasks, as it enables faster execution speeds and offers potential applications in areas like all-optical image analysis, feature detection, and object classification. (NA?)

  • Utilise deep generative models in order to effectively navigate the vast chemical space and identify optimal molecular structures for specific functionalities. (NA?)

  • Consider using weighted atom-centered symmetry functions (wACSFs) as descriptors in machine learning potentials, as they require fewer descriptors than traditional atom-centered symmetry functions (ACSFs) to achieve comparable spatial resolution, leading to improved generalization performance and reduced computational costs. (NA?)

  • Consider using a dynamic programming approach to calculate the edit-distance between layers in neural networks, while also accounting for skip-connections through a bipartite graph matching problem solved by the Hungarian algorithm. (NA?)

  • Consider utilizing advanced computational tools such as machine learning and deep learning algorithms alongside traditional medical imaging techniques like CT and MRI to improve the accuracy and efficiency of brain tumor diagnosis and classification. (NA?)

  • Consider using a translation-based methodology instead of a reconstruction-based methodology when developing molecular descriptors, as it forces the model to encode all necessary information of a given molecular representation into a compact latent space, leading to improved predictive performance in QSAR and virtual screening tasks. (NA?)

  • Utilize deep neural networks due to your capacity to efficiently capture complex functions and approximate any continuous function to any desired level of precision by allowing a sufficient number of units in a single hidden layer. (NA?)

  • Explore alternatives to traditional convolutional neural networks (CNNs) and transformers, such as the proposed MLP-Mixer architecture, which utilizes multi-layer perceptrons (MLPs) for both channel-mixing and token-mixing operations, resulting in competitive performance on image classification tasks. (NA?)

  • Carefully choose the most appropriate machine learning approach for your specific use-case, considering factors like the type of material, kind of data involved, spatial and temporal scales, formats, and desired knowledge gain, while balancing computational costs. (NA?)

  • Utilise deep neural networks (DNNs) for accurate predictions of chemical properties, specifically using the PhysNet architecture, which demonstrates superior performance across multiple benchmarks and effectively handles complexities such as long-range interactions and condensed phase systems. (NA?)

  • Consider implementing a Neuro-Fuzzy Inference System (WDT-ANFIS) based augmented wavelet de-noising technique for improving the accuracy of water quality predictions, particularly in cases where data might be affected by noise signals due to random and systematic errors. (NA?)

  • Carefully consider the choice of appropriate evaluation metrics when dealing with class imbalanced datasets, as common metrics like accuracy and error rate can be misleading in such scenarios. (NA?)

  • Employ scientometric analysis to evaluate global scientific production and development trends in the field of AI in health and medicine, providing insights into research gaps and informing policy development. (NA?)

  • Consider utilizing a novel approach called “deep 2BSDE method” when dealing with high-dimensional fully nonlinear partial differential equations (PDEs) and second-order backward stochastic differential equations (2BSDEs). This innovative technique combines a connection between PDEs and 2BSDEs, a merged formulation of the PDE and the 2BSDE problem, a temporal forward discretization of the 2BSDE and a spatial approximation via (NA?)

  • Consider using the Deep Learning Image Registration (DLIR) framework for unsupervised affine and deformable image registration, which trains ConvNets based on image similarity rather than requiring predefined example registrations, leading to increased efficiency and accuracy in medical imaging analysis. (NA?)

  • Consider integrating machine learning approaches like deep neural networks with traditional quantum chemistry methods to improve the accuracy and efficiency of molecular wavefunction predictions, leading to better understanding and optimization of molecular structures and properties. (NA?)

  • Use transfer learning to train a neural network on a large dataset of lower-accuracy DFT data, followed by retraining on a smaller dataset of higher-accuracy CCSD (T)/CBS data, to achieve a general-purpose potential that is both accurate and scalable across a variety of chemical systems. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Consider implementing Long Short-Term Memory (LSTM) networks in memristor crossbars to overcome limitations in computing power due to limited memory capacity and data communication bandwidth, thereby enhancing the potential of these networks for use in edge inference. (NA?)

  • Carefully evaluate the tradeoff between the complexity of your models and the quality of your data when selecting appropriate methods for analyzing your data. (NA?)

  • Develop and implement robust lifelong learning strategies for artificial learning systems, drawing inspiration from biological factors like structural plasticity, memory replay, curriculum and transfer learning, intrinsic motivation, and multisensory integration. (NA?)

  • Consider utilising all optical neural networks (AONNs) for machine learning tasks, as they offer the benefits of parallelism, low energy consumption, and scalability compared to traditional electronic-based methods. (NA?)

  • Focus on developing interactive refinement tools for users to communicate your preferences regarding the types of similarity that are most important at different moments in time, thereby increasing the diagnostic utility of images found and building user trust in the algorithm. (NA?)

  • Consider implementing lambda layers in your neural network architectures, as they enable efficient modeling of long-range interactions between input and structured contextual information, leading to improved performance and computational efficiency compared to traditional convolutional and attentional approaches. (NA?)

  • Focus on developing data-driven subgrid-scale models for partial differential equations (PDEs) using machine learning algorithms, specifically neural networks, to capture unresolved physics and improve the accuracy of numerical simulations. (NA?)

  • Focus on studying generalization of neural networks on small algorithmically generated datasets, as they offer a unique opportunity to examine data efficiency, memorization, generalization, and speed of learning in depth. (NA?)

  • Utilize tensor networks for machine learning tasks due to your potential for scalability, adaptability to both classical and quantum computing environments, and robust theoretical foundation. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Carefully choose appropriate machine learning algorithms, parallelism strategies, and system topologies to maximize the effectiveness and efficiency of your distributed machine learning systems. (NA?)

  • Consider implementing a concurrent learning approach for generating reliable deep learning-based potential energy surface (PES) models, which involves an interactive process of data generation and learning to ensure optimal representation and minimal size of the dataset. (NA?)

  • Consider the entire machine learning pipeline when developing visual analytics techniques, focusing on improving data quality and feature selection before model building, enhancing model understanding and diagnostics during model building, and supporting data interpretation after model building. (NA?)

  • Consider incorporating the Real-World-Weight Cross-Entropy (RWWCE) loss function into your machine learning models, especially when dealing with imbalanced classes or situations where the cost of mislabeling varies significantly among different categories. (NA?)

  • Consider utilizing the Contrastive Representation Learning (CRL) framework for developing and analyzing contrastive learning methods, as it provides a simplified and unified approach applicable to diverse data domains, learning setups, and definitions of similarity. (NA?)

  • Consider utilizing multimodal representation learning techniques, specifically focusing on the combination of vision and natural language modalities, to effectively integrate and process diverse forms of data in artificial intelligence applications. (NA?)

  • Utilise deep learning techniques for defect detection in manufacturing, taking into account various factors like the nature of the defect, the material being examined, and the specific requirements of the task. (NA?)

  • Utilise a combination of different computational methods to tackle the challenging task of drug screening and design, taking advantage of the strengths of each method to address issues at different scales and dimensions. (NA?)

  • Combine local eligibility traces and top-down learning signals in a specific way to create an effective online gradient descent learning method for recurrent spiking neural networks, called e-prop, which can approach the performance of backpropagation through time while remaining biologically plausible. (NA?)

  • Consider using committee machines, which involve combining multiple non-ideal memristor-based neural networks through ensemble averaging, to improve inference accuracy in physically implemented neural networks suffering from faulty devices, device-to-device variability, random telegraph noise, and line resistance. (NA?)

  • Use integrated gradients to optimize heatmaps for deep networks, as this approach leads to more accurate explanations of the networks decision-making processes compared to traditional gradient-based methods.’ (NA?)

  • Consider developing a compiler that converts floating-point machine learning models to fixed-point code for efficient deployment on Internet of Things (IoT) devices with limited memory resources. (NA?)

  • Carefully evaluate the stability of deep learning models in inverse problems, particularly in fields like medical imaging, as instabilities can lead to incorrect diagnoses and poor decision making. (NA?)

  • Use multiple analytical tools including the Pettitt test, Mann-Kendall (MK) test, Sens Innovative trend analysis, Artificial Neural Network-Multilayer Perceptron (ANN-MLP), and geostatistical techniques like Kriging in ArcGIS environment to comprehensively understand and forecast long-term Spatio-temporal changes in rainfall across different regions.’ (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Optimize the zero-shot learning objective directly by fine-tuning pre-trained language models on a collection of datasets, rather than relying solely on the next word prediction training objective. (NA?)

  • Incorporate specialized splicing scores into general variant effect prediction models to significantly enhance the accuracy of identifying pathogenic variants, while maintaining overall performance. (NA?)

  • Carefully evaluate the trade-off between stability and plasticity in continual learning algorithms, taking into account factors like model capacity, weight decay, and dropout regularization, and assessing performance across various benchmarks and datasets. (NA?)

  • Consider utilizing Bayesian Deep Learning (BDL) / Bayesian Neural Networks (BNNs) to enhance the reliability of your predictions, while addressing issues such as overfitting and providing valuable insights into the uncertainty of your models. (NA?)

  • Use a parallel algorithm for conservative PINNs (cPINNs) and extended PINNs (XPINNs) constructed with a hybrid programming model described by MPI + X, where X e {CPUs, GPUs}, to optimize all hyperparameters of each neural network separately in each subdomain, leading to improved performance for multi-scale and multi-physics problems. (NA?)

  • Consider utilizing automated machine learning (AutoML) tools throughout the entire machine learning pipeline, including data preparation, feature engineering, model generation, and model evaluation, to optimize model performance and minimize human intervention. (NA?)

  • Consider the interplay between cognitive barriers, digital routines, and organizational forms when investigating digital transformation in the modern competitive landscape. (NA?)

  • Utilize artificial intelligence (AI) and machine learning (ML) algorithms to enhance the drug discovery and development process, particularly in areas such as target identification, drug screening, and lead compound optimization, thereby reducing costs and time consumption. (NA?)

  • Consider employing complex-valued neural networks in optical computing systems, as they provide superior performance in terms of accuracy, convergence time, and construction of nonlinear decision boundaries compared to traditional real-valued neural networks. (NA?)

  • Consider developing a reconfigurable diffractive processing unit (DPU) for large-scale neuromorphic optoelectronic computing, which can be programmed to change its functionality and adapt to different types of neural network architectures, thereby significantly improving computing speed and system energy efficiency compared to existing electronic neuromorphic processors. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Utilise DeepONets, a novel neural network architecture consisting of two sub-networks - one for encoding the input function at a fixed number of sensors and another for encoding the locations for the output functions - to learn operators accurately and efficiently from a relatively small dataset. (NA?)

  • Utilise Relational Neural Descriptor Fields (R-NDFs) to efficiently and effectively determine the relative positioning of objects in space, even when dealing with previously unseen objects in varying positions. (NA?)

  • Consider using Binary Neural Networks (BNNs) for your projects, as they offer significant reductions in storage complexity and energy consumption compared to traditional neural networks, making them ideal for mobile and ultra-low power applications. (NA?)

  • Develop a prompt-based Chinese text classification framework that includes an automatic prompt generation process and an advanced candidate filtering method using mutual information and cosine similarity to enhance the performance of few-shot learning tasks. (NA?)

  • Consider using the Context Optimization (CoOp) technique when working with pre-trained vision-language models, as it enables automatic optimization of prompts for improved performance and reduced need for manual tuning. (NA?)

  • Make the most of free-text supervision when working with paired image and text data in the biomedical domain, particularly through careful text modelling, language grounding, augmentation, and regularization. (NA?)

  • Consider using Newtonian blurring as a novel approach to augmenting non-image biological datasets like human braingraphs, thereby enabling improved AI performance through increased sample sizes without introducing artificial alterations. (NA?)

  • Consider using a combination of dialogue classification and dialogue summarization methods, such as Support Vector Machines (SVM) and Graph Neural Networks (GNNs) for classification, and sequence-to-sequence (seq2seq) models with Recurrent Neural Networks (RNNs) or Transformer architectures for summarization, to efficiently process and analyze large amounts of medical text data. (NA?)

  • Utilize ControlNet, a neural network architecture designed specifically to add spatial conditioning controls to large, pretrained text-to-image diffusion models. This architecture effectively locks the production-ready large diffusion models and reuses your deep and robust encoding layers pretrained with billions of images as a strong backbone to learn a diverse set of conditional controls. By doing so, researchers can ensure that no harmful noise affects the fine-tuning process, thereby enabling more (NA?)

  • Consider incorporating essential matching signals, such as exact matching signals, semantic matching signals, and inference matching signals, into your analysis to enhance the generalizability of your findings across different domains and tasks. (NA?)

  • Investigate connectionist networks, focusing on developing efficient learning procedures that enable these networks to construct complex internal representations of your environment, while addressing challenges related to improving convergence rates and generalization abilities for application to larger, more realistic tasks. (NA?)

  • Explore the possibility of interpreting continuous prompts as a combination of discrete prompts, which could enhance the interpretability and transferability of continuous prompts in natural language processing tasks. (NA?)

  • Focus on developing efficient and unified neural architecture search frameworks, such as DDPNAS, which enable accurate and efficient searches across diverse search spaces and constraints. (NA?)

  • Consider employing ChatGPT as a valuable tool for debugging computer code, given its advanced natural language processing capabilities, extensive knowledge base, pattern recognition abilities, error correction capacity, and generalization power; however, the effectiveness of using ChatGPT for debugging depends on factors such as the specific task, the quality of the training data, and the design of the system. (NA?)

  • Utilise delta-tuning techniques to optimize large pre-trained language models (PLMs) for specific downstream tasks, thereby reducing computational costs without compromising performance. (NA?)

  • Combine the predictive power of AI with human expertise to optimize and accelerate the drug discovery process. (NA?)

  • Consider using an affinity scoring function to predict task transferability between pretrained language models, as it can efficiently identify beneficial tasks for transfer learning and reduce computational and storage costs compared to brute-force searches. (NA?)

  • Carefully consider your experimental setup to ensure validity and reliability in drawing conclusions about cause-and-effect relationships. (NA?)

  • Focus on collecting comprehensive and accurate data, ensuring it is free from artifacts and homogeneous, to effectively train AI algorithms and reduce inter- and intraobserver variability in CTG interpretation. (NA?)

  • Focus on developing deep learning algorithms that enable the discovery of increasingly abstract features within hierarchical representations, thereby promoting feature reuse and enhancing the overall effectiveness of machine learning systems. (NA?)

Artificial Neural Networks (Ann)

  • Carefully choose the depth and width of your deep neural networks to achieve the desired convergence rate in terms of number of training samples when applying the deep Ritz method (DRM) to solve partial differential equations (PDEs). (Y. Jiao et al. 2021)

  • Carefully consider the conflation of time and feature domains when developing saliency methods for time series data, and potentially adopt the proposed two-step temporal saliency rescaling (TSR) approach to improve the quality of saliency maps. (Ismail et al. 2020)

  • Carefully consider the trade-offs between the width and depth of artificial neural networks when attempting to learn complex boolean formulas, as this balance can significantly affect the efficiency and effectiveness of the learning process. (Nicolau et al. 2020)

  • Focus on developing models that can generalize well to new routes and cities, even if they dont have access to extensive training data. (Barnes et al. 2020)

  • Develop a systematic taxonomy of clustering methods that utilize deep neural networks, allowing them to create new clustering methods by selectively combining and modifying components of previous methods to overcome your individual limitations. (Aljalbout et al. 2018)

  • Consider combining domain alignment and discriminative feature learning when conducting unsupervised deep domain adaptation studies. (Yukang Chen et al. 2018)

  • Utilise MixGen, a novel multi-modal joint data augmentation approach, to significantly boost the efficiency and efficacy of your vision-language pre-training models. (Coulombe 2018)

  • Explore the potential of utilizing the Lottery Ticket Hypothesis (LTH) to identify fair and accurate subnetworks within densely connected neural networks, thereby reducing computational complexity while maintaining performance standards. (Frankle and Carbin 2018)

  • Carefully consider the potential impact of variance shift’ when combining batch normalization (BN) and dropout techniques in deep learning models, as this phenomenon can lead to numerical instability and reduced performance.’ (Xiang Li et al. 2018)

  • Develop a dynamic instance-specific threshold strategy for learning from noisy labels, allowing for improved identification and handling of varying levels of label noise within datasets. (W. Li et al. 2017)

  • Carefully evaluate the suitability of advanced machine learning techniques like Field-aware Factorization Machines (FFM) for real-world applications, considering aspects such as training time, memory requirements, and latency, and explore strategies to optimize these factors for practical deployment. (Juan, Lefortier, and Chapelle 2017)

  • Integrate the principles of efficient coding and Bayesian inference to create a comprehensive model of perceptual behavior, allowing them to better understand and predict various perceptual phenomena. (X.-X. Wei and Stocker 2015)

  • Consider using Unified DNAS (UDC) for generating state-of-the-art compressible neural networks (NNs) for NPU, which explores a large search space to balance trade-offs and improve performance. (Russakovsky et al. 2015)

  • Utilise deep neural networks (DNNs) to decode and predict neural responses to naturalistic stimuli, thereby revealing a gradient in the complexity of neural representations across the ventral stream. (Guclu and Gerven 2015)

  • Consider using a sequential inference framework for deep Gaussian processes (DGPs) to enable efficient processing of input-output data pairs, leading to improved performance and reduced computational costs. (Hensman and Lawrence 2014)

  • Consider incorporating selective classification techniques into your deep neural network models to improve prediction performance by trading off coverage, allowing users to set a desired risk level and maintain high levels of accuracy. (Simonyan and Zisserman 2014)

  • Consider using the Elastic Averaging Stochastic Gradient Descent (EASGD) algorithm for deep learning tasks in parallel computing environments, as it enables better exploration and improves overall performance compared to traditional methods like Downpour and ADMM. (Sixin Zhang, Choromanska, and LeCun 2014)

  • Aim to minimise the accuracy degradation associated with binarising convolutional neural networks (CNNs) by approximating full-precision weights with the linear combination of multiple binary weight bases and employing multiple binary activations. (Yoshua Bengio, Léonard, and Courville 2013)

  • Leverage a language model like GPT-3 to define a large space of possible bottlenecks, and then search for the best ones using a novel submodular utility that promotes the selection of discriminative and diverse information. (F. Bach 2010)

  • Apply the Bayesian model comparison framework to feedforward networks, enabling objective comparisons between different network architectures, choosing appropriate weight decay terms, estimating error bars on network parameters and output, and generating a measure of the effective number of parameters determined by the data. (Rossi and Vila 2006)

  • Consider developing a nonlinear model for neuronal interaction, which can become more linear at each successive stage of probabilistic analysis, leading to a better understanding of the complex dynamics underlying neuronal networks. (Sejnowski 1977)

  • Use symbiotic evolution in reinforcement learning models to promote cooperation and specialization among neurons, leading to faster and more efficient genetic search and avoiding convergence to suboptimal solutions. (NA?)

  • Focus on developing a constructive algorithm for training cooperative neural network ensembles (CNNe) that balances both accuracy and diversity among individual neural networks (NNs) in an ensemble, utilizing negative correlation learning and varying training epochs for individual NNs to enhance overall ensemble performance. (NA?)

  • Use a combination of regular expressions and machine learning techniques like neural networks to improve the accuracy of your predictions regarding Tat signal peptides, especially when dealing with variant forms that dont strictly adhere to the consensus pattern. (NA?)

  • Focus on developing machine learning algorithms that can effectively classify internet traffic without requiring access to sensitive information such as IP addresses or port numbers, thereby enhancing privacy protection while maintaining high levels of accuracy. (NA?)

  • Be cautious about your choice of protein samples when conducting studies involving machine learning programs, ensuring they are truly non-homologous to avoid potential biases in predictions. (NA?)

  • Carefully consider the choice of input variables when developing artificial neural networks (ANNs), as it affects model performance, computational effort, training difficulty, dimensionality, and comprehensibility. (NA?)

  • Consider utilizing hybrid Hidden Markov Model (HMM)/Artificial Neural Network (ANN) models for recognizing unconstrained offline handwritten texts, where the structural part of the optical models is modeled with Markov chains, and a Multilayer Perceptron is employed to estimate the emission probabilities. (NA?)

  • Consider adopting metaheuristic algorithms, such as evolutionary algorithms and swarm intelligence, alongside traditional gradient-based optimization methods, to overcome the limitations of these methods and enhance the generalization ability of feedforward neural networks. (NA?)

  • Consider developing and extending efficient and high-performance deep spiking neural networks (SNNs), focusing on your architectures and learning approaches, to better understand neural computation and different coding strategies in the brain, while potentially improving your performance on various tasks. (NA?)

  • Carefully consider your experimental setup, control for potential confounding variables, use appropriate statistical methods to analyze data, and interpret results with caution when drawing conclusions about causality. (NA?)

Convolutional Neural Networks (Cnn)

  • Consider using prompt tuning methods for speaker-adaptive visual speech recognition, specifically fine-tuning prompts on adaptation data of target speakers rather than modifying pre-trained model parameters, leading to significant improvements in performance for unseen speakers with minimal amounts of adaptation data. (Minsu Kim, Kim, and Ro 2023)

  • Consider implementing the Interventional Bag Multi-Instance Learning (IBMIL) technique to address the potential bias caused by the bag contextual prior in multi-instance learning (MIL) applications involving whole-slide pathological images (WSIs). (T. Lin et al. 2023)

  • Consider using a reparameterization encoder to optimize the generalizability of learnable prompts in vision-language models, improving your performance on unseen classes while maintaining your capacity to learn base classes. (Minh, Nguyen, and Tzimiropoulos 2023)

  • Utilise Equiangular Basis Vectors (EBVs) instead of the standard fully connected layer with softmax in deep neural networks for classification tasks. These EBVs predefine fixed normalised vector embeddings for each category, ensuring that the trainable parameters of the network remain constant even as the number of categories increases. This results in improved prediction accuracy and reduced computational costs. (Yang Shen, Sun, and Wei 2023)

  • Utilize the DeepMAD framework to design high-performance CNN models in a principled manner, leveraging constrained mathematical programming problems to optimize structural parameters without needing GPU or training data. (Xuan Shen et al. 2023)

  • Consider using an Object-Aware Distillation Pyramid (OADP) framework for open-vocabulary object detection, which involves an Object-Aware Knowledge Extraction (OAKE) module and a Distillation Pyramid (DP) mechanism to improve knowledge extraction and transfer efficiency. (Luting Wang et al. 2023)

  • Use the Knowledge-guided Context Optimization (KgCoOp) approach when working with visual-language models, as it helps reduce the discrepancy between learnable and hand-crafted prompts, thereby increasing the generalization ability of these models for unseen classes. (Hantao Yao, Zhang, and Xu 2023)

  • Consider utilizing a Dual Information Flow Network (DIFNet) to improve the accuracy of image captioning systems by incorporating segmentation features alongside traditional grid features, allowing for better integration of visual information and improved overall performance. (M. Wu et al. 2022)

  • Consider using a generative adversarial network (GAN)-like framework called GAN-MAE for your self-supervised learning tasks, as it offers significant computational efficiency and performance improvements over traditional masked autoencoder (MAE) techniques. (Assran et al. 2022)

  • Consider implementing a two-stage human activity recognition system on microcontrollers, utilizing a combination of decision trees and convolutional neural networks, to achieve improved energy efficiency without sacrificing accuracy. (Daghero, Pagliari, and Poncino 2022)

  • Consider incorporating learnable memory tokens into your Vision Transformer models to enhance your adaptability to new tasks while minimizing parameter usage and potentially preserving your capabilities on previously learned tasks. (Sandler et al. 2022)

  • Consider utilizing pre-trained deep learning models like ECAPA-TDNN and Wav2Vec2.0 to generate speech embeddings when working with limited datasets in stuttering detection tasks. (S. A. Sheikh et al. 2022)

  • Consider using Multiway Transformers for general-purpose modeling, enabling both deep fusion and modality-specific encoding, and performing masked “language” modeling on images, texts, and image-text pairs in a unified manner to achieve excellent transfer performance on both vision and vision-language tasks. (Wenhui Wang et al. 2022)

  • Focus on developing a comprehensive algorithm-circuit co-design framework that considers the unique characteristics of the target application and hardware constraints, allowing them to optimize the performance of your system while minimizing energy consumption and maximizing efficiency. (Datta et al. 2022)

  • Carefully consider the impact of data scaling on masked image modeling (MIM) performance, as MIM requires large-scale data to effectively scale up computes and model parameters, but cannot benefit from more data under a non-overfitting scenario. (H. Bao et al. 2021)

  • Employ data generators and distributed training techniques to overcome memory limitations and impracticably large training times when dealing with large neural networks and extensive seismic datasets. (Birnie, Jarraya, and Hansteen 2021)

  • Consider utilizing the sharpness-aware minimizer (SAM) optimizer to enhance the generalization capability of convolution-free architectures like ViTs and MLPs, thereby improving your overall performance. (Xiangning Chen, Hsieh, and Gong 2021)

  • Focus on developing a few-shot segmentation method based on dense Gaussian processes (GP) regression, which enables the capture of complex appearance distributions and provides a principled means of capturing uncertainty, leading to improved segmentation quality and robust cross-dataset transfer. (Johnander et al. 2021)

  • Consider implementing a shunted self-attention (SSA) technique in your Vision Transformer (ViT) models to enable the simultaneous modelling of both coarse-grained and fine-grained features, improving the models ability to handle images containing multiple objects of varying scales.’ (Sucheng Ren et al. 2021)

  • Consider incorporating heat diffusion methods into your transformer models when working with 3D mesh inputs, as it enables the model to adaptively capture multi-scale features and geometric structures, ultimately improving the overall performance of the model. (Yifan Xu et al. 2021)

  • Consider using convolutional neural networks (CNNs) as a tool for evaluating and comparing the performance of different classifications of elementary cellular automata (ECAs), since CNNs can effectively learn the underlying logic of these classifications and provide insightful comparisons based on your predictive accuracy. (Comelli, Pinel, and Bouvry 2021)

  • Utilize machine-driven design exploration strategies to develop highly efficient deep convolutional autoencoder network architectures for on-device acoustic anomaly detection, balancing accuracy and efficiency. (Müller et al. 2021)

  • Consider using deep learning techniques like convolutional neural networks (CNNs) for the accurate detection and classification of Ki-67 and tumor-infiltrating lymphocytes (TILs) in breast cancer, given the potential benefits of these methods in terms of speed, precision, and ability to learn optimal features from input data. (Negahbani et al. 2021)

  • Prioritize the development of deep learning architectures that facilitate the dense simultaneous modeling of multiresolution representation, as this significantly enhances the performance of tasks involving high-resolution dense prediction. (Sverrisson et al. 2020)

  • Consider combining convolutional neural networks (CNNs) and transformers to effectively model both local and global dependencies for image classification in an efficient manner. (Beyer et al. 2020)

  • Focus on reducing the size of intermediate activations required by back-propagation, instead of just focusing on reducing the number of trainable parameters, in order to effectively save training memory for efficient on-device learning. (Han Cai et al. 2020)

  • Integrate time series decomposition with deep neural networks for time series anomaly detection, as doing so allows for simpler network structures, improved model performance, and a more generalizable framework across various time series characteristics. (Jingkun Gao et al. 2020)

  • Carefully consider the choice of activation functions when building deep neural networks, as different types can lead to varying levels of model performance. (J. Heaton 2020)

  • Consider the directional inductive bias of neural networks when developing novel architectures, as it can significantly impact the performance and generalization capabilities of the models. (Ortiz-Jimenez et al. 2020)

  • Utilise Convolutional Occupancy Networks for 3D reconstruction tasks because it combines the advantages of convolutional neural networks and implicit representations, allowing for more accurate and scalable 3D reconstruction. (Songyou Peng et al. 2020)

  • Consider utilizing deep neural networks (DNNs) for weather forecasting tasks, particularly those involving precipitation, due to your ability to handle large spatial and temporal contexts, provide probabilistic outputs representing uncertainty, and adapt easily to increasing amounts of training data. (Sønderby et al. 2020)

  • Adopt a systematic evaluation and statistical analysis approach to ensure the validity and reliability of your results, particularly in the field of deep learning and computer vision. (Lathuiliere et al. 2020)

  • Consider integrating future data into model training for session-based recommendation systems, despite the challenge of avoiding data leakage, as it provides valuable signals about user preferences and can enhance recommendation quality. (F. Yuan et al. 2020)

  • Utilise Bayesian Optimisation to identify the ideal model architecture for Convolutional Neural Networks (CNNs) in order to achieve the highest performance levels. (Duong 2019)

  • Employ a three-stage process when balancing accuracy and sparsity in network training for keyword spotting tasks using convolutional neural networks (CNNs). (Sheen and Lyu 2019)

  • Consider implementing an Efficient Channel Attention (ECA) module when working with deep convolutional neural networks (CNNs), as it offers improved performance while reducing model complexity through avoiding dimensionality reduction and utilizing local cross-channel interactions. (Bello et al. 2019)

  • Consider both distribution-level and instance-level label matching issues when developing semi-supervised object detection systems, and propose solutions like re-distribution mean teachers and proposal self-assignments to mitigate these issues. (Kai Chen et al. 2019)

  • Utilise the Virtual Pooling (ViP) technique to enhance the efficiency of Convolutional Neural Networks (CNNs) in image classification and object detection tasks, thereby improving speed and energy consumption without significantly compromising accuracy. (Zhuo Chen et al. 2019)

  • Consider using a multi-granularity contrasting (MGC) framework when working on cross-lingual pre-training tasks, as it combines the benefits of bidirectional context modeling and embedding alignment, leading to improved performance in various downstream tasks such as machine translation and cross-lingual language understanding. (Chi et al. 2019)

  • Consider using diffusion transformers (DiTs) as a replacement for the conventional U-Net backbone in diffusion models due to your superior scalability properties and potential benefits from architecture unification. (Child et al. 2019)

  • Consider using diverse datasets and employing various techniques such as heavy augmentation of training data, network regularization, and margin penalties to avoid overfitting and achieve better performance in speaker recognition tasks. (J. S. Chung et al. 2019)

  • Consider using parameterized convolutional neural networks (PCNNs) for aspect level sentiment classification, as demonstrated by the authors successful implementation of PCNNs achieving state-of-the-art results on SemEval 2014 datasets.’ (B. Huang and Carley 2019)

  • Consider using pretrained audio neural networks (PANNs) trained on large-scale datasets like AudioSet for improved performance in audio pattern recognition tasks, while exploring the trade-offs between performance and computational complexity. (Q. Kong et al. 2019)

  • Consider using the Rectified Local Phase Volume (ReLPV) block as an efficient alternative to the traditional 3D convolutional layer in 3D CNNs, as it offers significant parameter savings, improved feature learning capabilities, and consistent performance improvements across different 3D data representations. (Kumawat and Raman 2019)

  • Utilise structured sparsity regularisation (SSR) when working with convolutional neural networks (CNNs) to achieve simultaneous computational speed up and memory overhead reduction. This approach involves incorporating two types of structured sparsity regularisers into the original objective function of filter pruning, allowing for the coordination of global outputs and local pruning operations to adaptively prune filters. Furthermore, it proposes an Alternative Updating with Lagrange Multipliers ( (S. Lin et al. 2019)

  • Consider using basis point sets (BPS) as a highly efficient and fully general way to process point clouds with machine learning algorithms, as demonstrated by matching the performance of PointNet on a shape classification task while using three orders of magnitude fewer floating point operations. (Prokudin, Lassner, and Romero 2019)

  • Utilise a combination of deformable convolution (DCN) and transformer-style components within your convolutional neural networks (CNNs) to enable the CNNs to learn long-range dependencies and adaptive spatial aggregation, thereby improving your ability to handle large-scale datasets and compete with transformer-based models. (Shoeybi et al. 2019)

  • Focus on increasing feature interactions when developing convolution-based knowledge graph embeddings, as doing so improves link prediction performance. (Vashishth et al. 2019)

  • Use a min-entropy latent model (MELM) for weakly supervised object detection tasks, as it helps to reduce the variance of positive instances and alleviate the ambiguity of detectors. (Wan et al. 2019)

  • Consider applying Hessian-based structured pruning methods in the Kronecker-factored eigenbasis (KFE) rather than in parameter coordinates, as this approach enables accurate pruning and faster computation, particularly for more challenging datasets and networks. (Chaoqi Wang et al. 2019)

  • Consider incorporating external knowledge from law provisions and a suitable way to decide label numbers when developing models for legal charge prediction tasks. (D. Wei and Lin 2019)

  • Focus on developing models that combine across-task learning of the network and per-class reference vectors with quick task-adaptive conditioning of classification space, allowing for excellent generalization to new data. (S. W. Yoon, Seo, and Moon 2019)

  • Consider using Summed-Area Tables (SATs) and box filters to perform large-kernel convolution in fully-convolutional neural networks, allowing for efficient combination of high-resolution output with wide receptive fields for pixel-level prediction tasks. (Linguang Zhang, Halber, and Rusinkiewicz 2019)

  • Consider combining pruning and quantization techniques to achieve optimal compression of deep convolutional neural networks (CNNs) while maintaining high task accuracy. (Yiren Zhao et al. 2019)

  • Consider the network compression problem from a new perspective where the shape of the weight tensors and the architecture are designed independently, enabling the network parameters to be disentangled from the architecture and compactly represented by a small-sized parameter set (called epitome). (D. Zhou et al. 2019)

  • Focus on optimizing the use of FPGAs as accelerators for deep learning networks by addressing implementation challenges related to storage, external memory bandwidth, and computational resources, while considering the unique characteristics of different layers in CNNs. (Shawahna, Sait, and El-Maleh 2019)

  • Carefully consider the structural constraints and external factors affecting the distribution of flows in a given region when developing models for fine-grained urban flow inference. (Yuxuan Liang et al. 2019)

  • Consider pre-training deep neural networks on multiple document datasets rather than solely on natural scene images to achieve improved performance in text line detection tasks. (Boillet et al. 2019)

  • Consider implementing a novel method called PruneTrain, which combines group lasso regularization with dynamic network reconfiguration to continuously prune and optimize the architecture of convolutional neural networks during training, thereby reducing computational, memory, and communication costs without compromising model accuracy. (Lym et al. 2019)

  • Consider using the PointGrid method when dealing with 3D shape understanding problems, as it offers superior performance over existing deep learning methods on both classification and segmentation tasks. (T. Le and Duan 2018)

  • Utilise convex optimisation methods to identify sparse sets of weights in deep neural networks, leveraging decades of research in convex optimization to achieve scalability and predictable convergence behaviour. (Aghasi, Abdi, and Romberg 2018)

  • Consider employing deep learning approaches, particularly convolutional neural networks (CNNs), recurrent neural networks (RNNs), and deep reinforcement learning (DRL), depending on the nature of the problem and availability of labeled data, to achieve state-of-the-art performance across various domains. (Alom, Taha, et al. 2018)

  • Employ the Inception Recurrent Residual Convolutional Neural Network (IRRCNN) model for breast cancer classification from histopathological images, as it demonstrates superior performance against equivalent Inception Networks, Residual Networks, and Recurrent Convolutional Neural Networks (RCNNs) for object recognition tasks. (Alom, Yakopcic, et al. 2018)

  • Focus on developing novel methods for accelerating and compressing convolutional layers in neural networks through filter quantization and clustering, rather than solely relying on tensor decomposition techniques. (Babin et al. 2018)

  • Leverage the well-understood and well-modeled structure of language, through classical NLP parsing and/or use of the modern pre-trained LLMs, for manipulating the text part of the standard VL paired datasets to regularize VL training and teach SVLC understanding to VL models. (Battaglia et al. 2018)

  • Pay careful attention to the choice of convolutional neural network architecture when working with self-supervised visual representation learning, as it can greatly impact the performance of the model. (Behrmann et al. 2018)

  • Employ reinforcement learning based on actor-critic structure to optimize the compression of deep neural networks, resulting in significant improvements in model compression quality without requiring human intervention. (Hakkak 2018)

  • Consider implementing a combination of training procedure refinements and model architecture tweaks to achieve significant improvements in model accuracy for image classification tasks, ultimately leading to better transfer learning performance in other application domains. (Tong He et al. 2018)

  • Utilize Partial Least Squares (PLS) and Variable Importance in Projection (VIP) to effectively identify and remove less significant filters in convolutional networks, leading to reduced computational costs without compromising network accuracy. (Jordao et al. 2018)

  • Utilise a combination of DNN partitioning and DNN right-sizing techniques to achieve low-latency edge intelligence, particularly for mission-critical applications like VR/AR games and robotics. (E. Li, Zhou, and Chen 2018)

  • Consider multiple factors beyond just final performance when evaluating the effectiveness of a pruning method for deep convolutional neural networks, including the initial drop in performance, the degree of recovery, the speed of recovery, and the quantity of data needed for recovery. (D. Mittal et al. 2018)

  • Consider utilizing a deep residual network of convolutional and recurrent units for earthquake signal detection, as demonstrated by the authors development of the Cnn-Rnn Earthquake Detector (CRED) which achieved impressive results in terms of sensitivity, robustness, and efficiency.’ (Mousavi et al. 2018)

  • Leverage the power of partial differential equations (PDEs) to analyze and optimize deep learning tasks, particularly in the areas of image processing and classification. (Ruthotto and Haber 2018)

  • Consider integrating competitive learning into your convolutional neural networks (CNNs) to enhance representation learning and increase the efficiency of fine-tuning, particularly when dealing with large amounts of unlabelled data. (Shinozaki 2018)

  • Consider using an incremental regularization approach for efficient ConvNets, which involves assigning different regularization factors to different weight groups based on your relative importance, allowing for a more gradual adaptation of the network during pruning. (Huan Wang et al. 2018)

  • Focus on developing a principled and effective method to model dynamic skeletons and leverage them for action recognition, moving beyond conventional approaches that rely on hand-crafted parts or traversal rules. (Sijie Yan, Xiong, and Lin 2018)

  • Consider the limitations of traditional regularization-based pruning techniques, particularly in terms of scalability and compatibility with batch normalization, and explore alternative approaches such as imposing sparsity on the scaling parameter γ in batch normalization operators to improve efficiency and accuracy in deep learning models. (J. Ye et al. 2018)

  • Use a recursive Bayesian pruning method (RBP) to efficiently prune channels in convolutional neural networks while considering inter-layer dependencies, leading to significant improvements in computational efficiency without sacrificing model accuracy. (Yuefu Zhou et al. 2018)

  • Consider combining multiple compression techniques, such as parameter pruning and sharing, low-rank factorization, transferred/compact convolutional filters, and knowledge distillation, to effectively reduce the size and computational requirements of deep neural networks while preserving your performance. (Y. Cheng et al. 2018)

  • Explore the potential benefits of using graph convolutional networks (GCNs) for text classification tasks, particularly when dealing with limited amounts of training data, as GCNs can effectively capture global word co-occurrences and lead to improved classification performance compared to traditional approaches. (Yifu Li, Jin, and Luo 2018)

  • Consider using a genetic algorithm (GA) for pruning convolutional neural networks (CNNs) based on a multi-objective trade-off between error, computation, and sparsity, as demonstrated through its successful application in reducing parameter size and improving computation efficiency while maintaining acceptable accuracy levels. (“Artificial Neural Networks and Machine Learning – ICANN 2018” 2018)

  • Utilise the learnable graph convolutional layer (LGCL) to enable the application of regular convolutional operations on graph data, rather than modifying the convolutional operations to suit the graph data. (H. Gao, Wang, and Ji 2018)

  • Consider integrating multiple information sources, including visual patterns, textual semantics, and presentation structures, when estimating the relevance of search results. This approach allows for a more accurate understanding of how users judge the relevance of search results, taking into account factors beyond just textual content. (Junqi Zhang et al. 2018)

  • Utilise sparse convolutional networks for LiDAR-based object detection to significantly increase the speed of both training and inference, whilst also improving orientation estimation performance through a new angle loss regression technique and enhancing convergence speed and performance through a novel data augmentation approach. (Yan Yan, Mao, and Li 2018)

  • Focus on utilizing deep learning methods for improved performance in acoustic scene classification, sound event detection, and domestic audio tagging tasks, while maintaining consistent feature representations across tasks. (Mesaros et al. 2017)

  • Consider incorporating the cutout regularization technique in your convolutional neural networks to improve model robustness and overall performance, especially when working with limited data or high-resolution images. (DeVries and Taylor 2017)

  • Utilize a fully convolutional architecture for sequence to sequence modeling instead of relying solely on recurrent neural networks, enabling improved performance on large-scale tasks while reducing computational complexity. (Gehring et al. 2017)

  • Utilise the Super Learner methodology when working with deep convolutional neural networks for image classification tasks, due to its ability to outperform other ensemble methods in terms of accuracy and adaptivity. (Ju, Bibaut, and Laan 2017)

  • Consider using a “Learning with Rethinking” algorithm, which involves adding a feedback layer and producing an emphasis vector to enable your convolutional neural network (CNN) models to recurrently boost performance based on previous predictions. (Xin Li et al. 2017)

  • Carefully consider the unique characteristics of IoT data when selecting and applying deep learning techniques for IoT big data and streaming analytics, taking into account factors such as data volume, velocity, variety, veracity, variability, and value. (Mohammadi et al. 2017)

  • Carefully examine and optimize your convolutional neural network architectures using a combination of qualitative and quantitative analysis techniques, such as confusion matrices, validation curves, learning curves, and input-feature based model explanations, while considering factors such as batch size, ensemble averaging, data augmentation, and test-time transformations to achieve improved performance. (Thoma 2017)

  • Aim to build statistical models that take into account any known symmetries in the underlying data, as doing so can greatly simplify the learning task and improve overall performance. (Weiler, Hamprecht, and Storath 2017)

  • Consider multiple factors beyond just final performance when evaluating the effectiveness of a pruning method, including the initial drop in performance, the degree of recovery, the speed of recovery, and the amount of data needed for recovery. (Francois Chollet 2017)

  • Aim to develop efficient convolution operators for spatial redundancy pruning, specifically through the use of a magnitude-based sampling module incorporated into 3D convolution layers to reduce redundancy in data and model. (J. Dai et al. 2017)

  • Consider combining deep learning networks with model-based methods to achieve superior performance in jointly reconstructing MR images and coil sensitivity maps from undersampled multi-coil k-space data. (Diamond et al. 2017)

  • Focus on developing a comprehensive understanding of the specific characteristics of legal texts, including your unique structure and terminology, in order to create effective information retrieval and question answering systems. (P.-K. Do et al. 2017)

  • Consider incorporating deep neural networks (DNNs) into your video delivery frameworks to enhance video quality independently of available bandwidth, thereby improving overall user quality of experience (QoE). (Hanzhang Hu et al. 2017)

  • Consider using a data-driven, end-to-end approach for selecting sparse structures in deep neural networks, rather than relying solely on expert knowledge or extensive experimentation. (Zehao Huang and Wang 2017a)

  • Consider implementing Binarized Convolutional Neural Networks with Separable Filters (BCNNw/SF) to achieve significant reductions in computational and storage complexity when working with large-scale neural networks. (J.-H. Lin et al. 2017)

  • Pay close attention to the selection of appropriate training data for speech emotion recognition systems, as the type of speech data used can greatly impact the overall performance of the system. (M. Neumann and Vu 2017)

  • Consider extending the Winograd algorithm to Residue Number System (RNS) for more efficient and accurate convolution in low-precision quantized neural networks. (Krizhevsky, Sutskever, and Hinton 2017)

  • Consider using flex-convolution, a natural generalization of traditional convolution layers, for processing unstructured data like 3D point clouds, as it offers competitive performance on small benchmark sets and significant improvements on million-scale real-world datasets, while requiring fewer parameters and lower memory consumption. (“Pattern Recognition” 2017)

  • Consider utilizing a three-stage pipeline incorporating convolutional neural networks (CNNs) to effectively identify Northern Leaf Blight (NLB)-infected maize plants from field imagery, thereby improving diagnostic accuracy and reducing the need for labor-intensive manual inspection. (DeChant et al. 2017)

  • Carefully consider the feasibility of mapping a given CNN computation onto a systolic array structure, taking into account factors such as data reuse, PE array shape, and data reuse strategy, in order to optimize system throughput and minimize resource consumption. (Xuechao Wei et al. 2017)

  • Consider using a combination of low-rank CP-decomposition with Tensor Power Method (TPM) for efficient optimization and iterative fine-tuning to overcome the instability issues associated with CP-decomposition in order to effectively compress convolutional neural networks (CNNs) for improved performance on resource-constrained devices. (Astrid and Lee 2017)

  • Consider utilizing low-rank tensor decomposition of convolutional weights to modify neural network architecture, incorporating sparsity-inducing regularizers to enable structured pruning, and combining light-weight neural networks with radial basis functions for rapid fine-grained classification, resulting in substantial speedups for contemporary convolutional architectures. (B. Baker et al. 2017)

  • Consider sharing convolutional layer weights within residual blocks operating at the same spatial scale to reduce the number of parameters required in deep residual networks without sacrificing significant accuracy. (Boulch 2017)

  • Consider implementing sparse connections in Convolutional Neural Networks (CNNs) to achieve better performance and efficiency, particularly in cases where dense convolutions may lead to redundancy and increased computational costs. (Changpinyo, Sandler, and Zhmoginov 2017)

  • Consider incorporating task identification information into your class-incremental learning algorithms, as it can lead to significant improvements in performance. (DeVries and Taylor 2017)

  • Consider using a simple hill climbing procedure with network morphisms and cosine annealing for efficient architecture search in convolutional neural networks, as it significantly reduces computational costs while maintaining competitive performance. (Elsken, Metzen, and Hutter 2017)

  • Model individual labelers instead of treating the majority opinion as the correct label or modelling the correct label as a distribution, allowing for improved classification results. (Guan et al. 2017)

  • Focus on developing a recurrent convolutional network for real-time video style transfer that incorporates a temporal consistency loss to improve the stability of existing methods. (A. Gupta et al. 2017)

  • Develop an iterative two-step algorithm for effective channel pruning in deep convolutional neural networks, involving LASSO regression-based channel selection and least square reconstruction, to reduce accumulated error and enhance compatibility across various architectures. (Yihui He, Zhang, and Sun 2017)

  • Focus on developing efficient network architectures like CondenseNet, which combine dense connectivity with learned group convolutions to optimize feature reuse while removing unnecessary connections, ultimately enabling faster and more efficient computations on mobile devices. (G. Huang et al. 2017)

  • Consider incorporating introspective convolutional networks (ICN) into your experimental designs, as these networks enable simultaneous generative and discriminative learning, leading to improved classification results. (L. Jin, Lazarow, and Tu 2017)

  • Use a soft product quantization layer within your neural networks to enable end-to-end training of the product quantization network, while employing an asymmetric triplet loss to optimize the asymmetric similarity measurement. (B. Klein and Wolf 2017)

  • Consider using end-to-end neural speaker embedding systems, such as Deep Speaker, which combine all three steps of traditional i-vector systems, optimize them jointly, and reduce the mismatch between training and test phases. (Chao Li et al. 2017)

  • Utilise the Winograd layer as an architectural component in your deep learning models. This allows for efficient pruning of Winograd parameters, leading to faster inference times without compromising accuracy. (Sheng Li, Park, and Tang 2017)

  • Consider implementing network slimming, a method that reduces model size, decreases runtime memory footprint, and lowers the number of computing operations in deep convolutional neural networks, without sacrificing accuracy. (Zhuang Liu et al. 2017)

  • Focus on filter level pruning for deep neural networks, specifically by evaluating the importance of each filter based on the outputs of its next layer rather than its own layer, allowing for simultaneous acceleration and compression of CNN models with minimal performance degradation. (J.-H. Luo, Wu, and Lin 2017)

  • Consider using coarse-grained pruning when working with deep neural networks, as it offers a balance between maintaining prediction accuracy and improving hardware efficiency through increased sparsity regularity. (Huizi Mao et al. 2017)

  • Consider using Two-Bit Networks (TBNs) for model compression of Convolutional Neural Networks (CNNs) on resource-constrained embedded devices, as it allows for reduced memory usage and improved computational efficiency while maintaining good classification accuracy. (Wenjia Meng et al. 2017)

  • Consider utilizing the Neural Side-By-Side methodology when comparing super-resolution models, as it provides an automatic and efficient way to approximate human preferences, thereby enabling accurate model comparison and hyperparameter tuning without requiring direct human intervention. (Murray and Gordo 2017)

  • Employ a fully-convolutional character-to-spectrogram architecture for speech synthesis, which enables fully parallel computation and trains significantly faster than analogous architectures using recurrent cells. (Ping et al. 2017)

  • Focus on developing dynamic network surgery techniques that involve both pruning and splicing operations to effectively compress deep neural networks without compromising your predictive accuracy. (Courbariaux et al. 2016)

  • Consider implementing the “Learning Without Forgetting” (LwF) method when attempting to add new capabilities to a Convolutional Neural Network (CNN) without access to the original training data, as it effectively preserves the original capabilities while allowing for the addition of new ones. (Ke Li and Malik 2016)

  • Explore the potential of deep learning algorithms for medical image reconstruction, particularly in situations where traditional methods struggle, due to your ability to learn from large amounts of data and perform powerful multi-scale analysis. (Jingdong Wang et al. 2016)

  • Consider integrating both pruning and hints techniques in your model acceleration frameworks, as they are complementary and can lead to improved performance. (Alvarez and Petersson 2016)

  • Use tensor factorization methods to compress convolutional layers in neural networks, achieving significant reductions in computational and memory complexity while maintaining comparable levels of accuracy. (Garipov et al. 2016)

  • Utilize a deep 3D convolutional neural network (3D-CNN) pretrained by a 3D Convolutional Autoencoder (3D-CAE) to learn generic discriminative AD features in the lower layers, which can be easily adapted to datasets collected in different domains, and enforce a discriminative loss function on upper layers (deep supervision) to increase the specificity of features. (Hosseini-Asl, Gimel’farb, and El-Baz 2016)

  • Consider developing a compact DNN architecture that utilises a new module called Conv-M’, which enables the extraction of diverse feature extractors without significantly increasing parameters, thus improving the overall performance of the DNN in both classification and domain adaptation tasks.’ (Iandola et al. 2016)

  • Consider pruning filters rather than individual weights in order to efficiently reduce computation costs in convolutional neural networks (CNNs) without compromising accuracy. (Hao Li et al. 2016)

  • Directly use energy consumption as a metric to guide the design of convolutional neural networks (CNNs) rather than focusing on the number of weights or operations, as this better aligns with the actual energy usage patterns of these networks. (T.-J. Yang, Chen, and Sze 2016)

  • Carefully consider the balance between resource utilization and accuracy when developing deep neural networks for continuous mobile vision applications, taking into account factors such as memory use, execution energy, and execution latency. (Seungyeop Han et al. 2016)

  • Consider utilising a deeply pipelined multi-FPGA architecture to expand the design space for optimal performance and energy efficiency in Convolutional Neural Network (CNN) applications. (Chen Zhang et al. 2016)

  • Consider integrating semantic relationships among fine-grained classes in your visual food recognition frameworks through the use of a multi-task loss function on top of a convolutional neural network (CNN) architecture, followed by a random walk based smoothing procedure to further exploit the rich semantic information. (H. Wu et al. 2016)

  • Consider incorporating multiple aspects of conversational context when developing models for predicting responses in open-domain, multi-turn, unstructured, multi-participant conversations, including both the immediate context of the preceding message and the broader historical context of the conversation and individual participants. (Al-Rfou et al. 2016)

  • Consider using deep convolutional neural networks (CNNs) for automated knee osteoarthritis (OA) severity assessment, as they demonstrated significant improvements in classification accuracy when compared to previous methods. Additionally, the authors suggest framing the prediction of KL grades as a regression problem, leading to even greater accuracy gains. (Antony et al. 2016)

  • Consider adopting the Multiplicative Fourier Level of Detail (MFLOD) technique for improved accuracy and scalability in implicit neural representation tasks, as it enables explicit bandwidth control for each level of detail and offers greater feasibility in Fourier analysis compared to traditional methods. (J. L. Ba, Kiros, and Hinton 2016)

  • Consider incorporating heterophily-aware mechanisms when working with complex visual scenes, as doing so can improve the accuracy of scene graph generation algorithms. (J. L. Ba, Kiros, and Hinton 2016)

  • Consider utilizing a lookup-based convolutional neural network (LCNN) for efficient learning and inference in resource-constrained environments, as it enables fast, compact, and accurate modeling by encoding convolutions via a few lookups to a trained dictionary. (Bagherinezhad, Rastegari, and Farhadi 2016)

  • Adopt a gradient-based architecture search with resource constraints for object detection tasks, using the proposed Auto-FPN framework that includes Auto-fusion and Auto-head modules to optimize feature fusion and classification/bounding-box regression respectively. (B. Baker et al. 2016)

  • Consider replacing traditional Inception modules with depthwise separable convolutions in neural computer vision architectures, as this approach offers improved efficiency and performance. (François Chollet 2016)

  • Consider using an end-to-end automatic speech recognition system that combines a standard 1D convolutional neural network, a sequence criterion which can infer the segmentation, and a simple beam-search decoder, as it offers competitive results on the LibriSpeech corpus with MFCC features (7.2% WER), and promising results with power spectrum and raw speech (9.4% WER and 10.1% WER respectively), (Collobert, Puhrsch, and Synnaeve 2016)

  • Focus on developing an exclusive feature map dimensionality reduction method for deep network compression problems, specifically by employing circulant matrices for projection to ensure low space complexity and high mapping speed. (Courbariaux et al. 2016)

  • Consider implementing a variational Bayesian scheme for pruning convolutional neural networks at the channel level, as it offers improvements in computation efficiency and stability compared to traditional deterministic value-based pruning methods. (Courbariaux et al. 2016)

  • Focus on developing dynamic network surgery techniques for efficient deep neural network compression, which involve both pruning and splicing operations to ensure accurate and efficient network maintenance. (Courbariaux et al. 2016)

  • Utilise the proposed temporal network-diffusion convolution networks’ (TNDCN) model for analysing dynamic social interaction networks. This model enables unified representation learning for multiple downstream tasks with minimal need for knowledge-based feature engineering, and has demonstrated superior performance in tasks such as deception, dominance, and nervousness detection.’ (H. Dai et al. 2016)

  • Focus on developing efficient High-Order DEcomposed Convolution (HODEC) techniques to simultaneously reduce computational and storage costs in deep neural networks, thus overcoming the computation inefficiency issue associated with traditional tensor decomposition approaches. (Garipov et al. 2016)

  • Consider using a convolutional encoder model for neural machine translation due to its ability to encode the source sentence simultaneously, leading to increased efficiency and competitive accuracy compared to recurrent networks. (Gehring et al. 2016)

  • Focus on developing hardware-oriented model approximation techniques, such as Ristretto, to optimize the efficiency of Convolutional Neural Networks (CNNs) by balancing bit-width reduction and accuracy loss, ultimately leading to faster and more efficient implementations. (Gysel, Motamedi, and Ghiasi 2016)

  • Consider designing smaller convolutional neural networks (CNNs) with fewer parameters, as they offer significant benefits in terms of efficiency, ease of deployment, and feasibility for use in resource-constrained environments like FPGAs and embedded systems, without compromising on accuracy. (Iandola et al. 2016)

  • Apply the Pruning in Training (PiT) framework when working with Deep Convolutional Neural Networks (DCNNs) to effectively reduce the parameter size while maintaining comparable performance. (K. Jia 2016)

  • Utilise a combination of graph convolution networks (GCN) and graph attention networks (cosAtt) within a spatial gated block to effectively capture complex spatial-temporal features in traffic prediction tasks. (Kipf and Welling 2016a)

  • Focus on developing a unified architecture for your convolutional neural network (CNN) that can handle various levels of vision tasks, including low-, mid-, and high-level tasks, while being trained end-to-end. The authors suggest that this approach can help overcome issues associated with training a deep architecture using diverse training sets and limited memory budgets, ultimately leading to improved overall performance. (Kokkinos 2016)

  • Utilise Convolutional Neural Networks (CNNs) for solving complex machine learning tasks, particularly those involving natural images, due to your ability to effectively handle local symmetries and translate variations in the input data. (Koushik 2016)

  • Consider combining multiple networks, each specialized for different phases of a complex task, to enhance overall performance. (Lample and Chaplot 2016)

  • Utilise logarithmic data representation when working with convolutional neural networks, as it allows for improved classification accuracy while reducing the precision needed for encoding weights and activations. (Miyashita, Lee, and Murmann 2016)

  • Consider using PointNet, a novel deep learning architecture that directly consumes point clouds, rather than converting them to regular 3D voxel grids or collections of images, as it respects the permutation invariance of points in the input and offers a unified architecture for applications ranging from object classification, part segmentation, to scene semantic parsing. (C. R. Qi et al. 2016)

  • Consider using Product-based Neural Networks (PNNs) when attempting to predict user responses, as they offer improved performance compared to existing methods due to your ability to capture interactive patterns between inter-field categories and explore high-order feature interactions. (Y. Qu et al. 2016)

  • Consider using deep convolutional neural networks for image classification tasks, as they outperform shallow models even when trained to mimic the latter. (Urban et al. 2016)

  • Focus on developing efficient convolutional layers through techniques like single intra-channel convolution, topological subdivisioning, and spatial “bottleneck” structure to optimize the accuracy/complexity ratio in deep convolutional neural networks. (Min Wang, Liu, and Foroosh 2016)

  • Consider using deep neural networks for end-to-end time series classification without any heavy preprocessing or feature engineering, as they offer comparable or even superior performance compared to traditional methods. (Zhiguang Wang, Yan, and Oates 2016)

  • Utilize convolutional and LSTM neural networks, along with a novel spatial smoothing method and lattice-free MMI acoustic training, to achieve human parity in conversational speech recognition. (W. Xiong et al. 2016)

  • Focus on developing efficient algorithms for training low bitwidth neural networks using low bitwidth gradients, enabling faster training times and lower memory requirements without sacrificing prediction accuracy. (S. Zhou et al. 2016)

  • Leverage the strengths of Convolutional Neural Networks (CNNs) in handling image-based problems, while paying attention to potential issues such as overfitting and computational complexity, and applying appropriate strategies such as parameter sharing and pooling layers to optimize the performance of the network. (K. O’Shea and Nash 2015)

  • Consider using deep convolutional neural networks (DCNNs) for feature extraction in your studies, as these networks provide translation invariance and limited sensitivity to deformations, leading to improved classification performance. (Wiatowski and Bölcskei 2015)

  • Consider utilizing a gradient descent-based approach for architecture compression, which involves encoding an input architecture into a continuous latent space and performing gradient descent on the encoded feature to optimize a compression objective function that balances accuracy and parameter count. (Girshick 2015)

  • Carefully choose your baseline, model parameters, and hardware when exploring the benefits of ultra-low-precision models in mobile computer vision applications. (Zee and Geijn 2015)

  • Consider using cross-image-attention for conditional embeddings in deep metric learning to improve the accuracy of your models. (Jian Guo and Gould 2015)

  • Explore various deep neural network architectures to combine image information across a video over longer time periods than previously attempted, considering both convolutional temporal feature pooling architectures and recurrent neural networks that use Long Short-Term Memory (LSTM) cells. (Ng et al. 2015)

  • Consider using data augmentation techniques like elastic deformations to improve the efficiency of your training process, allowing them to work effectively with fewer annotated samples. (Ronneberger, Fischer, and Brox 2015)

  • Utilise a unified framework called “Quantized CNN” to simultaneously accelerate and compress convolutional networks, thereby enabling faster test-phase computations and reducing storage and memory consumption. (Jiaxiang Wu et al. 2015)

  • Utilise a convolutional neural network to create continuous representations for textual relations, thereby enhancing overall performance on link prediction tasks, especially for entity pairs that have textual mentions. (Toutanova et al. 2015)

  • Consider utilizing a Convolutional Click Prediction Model (CCPM) for click prediction in scenarios involving single ad impressions and sequential ad impressions, as it effectively mines significant semantic features through convolutional layers and dynamic pooling layers, leading to improved accuracy in click prediction. (Qiang Liu et al. 2015)

  • Develop an iterative two-step algorithm for effective channel pruning in deep convolutional neural networks, involving LASSO regression-based channel selection and least square reconstruction, to reduce accumulated error and enhance compatibility across various architectures. (Anwar, Hwang, and Sung 2015)

  • Utilise SpiderCNN, a new convolutional architecture specifically designed for direct extraction of features from point clouds, rather than relying on traditional convolutional neural networks (CNNs) which struggle with the irregular distribution of point clouds in R^3. (A. X. Chang et al. 2015)

  • Utilise a data-driven point cloud upsampling technique that learns multi-level features per point and expands the point set via a multi-branch convolution unit implicitly in feature space. (A. X. Chang et al. 2015)

  • Focus on developing and comparing various deep learning architectures for improving the performance of non-factoid question answering tasks, such as through the use of convolutional neural networks (CNNs) and different similarity metrics. (M. Feng et al. 2015)

  • Focus on developing methods to efficiently identify and eliminate unnecessary connections in neural networks, thereby improving overall network performance and reducing computational costs. (Song Han et al. 2015)

  • Consider implementing a one-shot whole network compression scheme when working with deep convolutional neural networks for fast and low power mobile applications. (Y.-D. Kim et al. 2015)

  • Consider adding global context to your fully convolutional networks for semantic segmentation, as it can lead to significant improvements in accuracy with minimal computational overhead. (Wei Liu, Rabinovich, and Berg 2015)

  • Utilize low-rank tensor decompositions to simplify and improve deep convolutional neural networks (CNNs) for faster processing and potentially improved performance. (C. Tai et al. 2015)

  • Explore the use of convolutional neural networks (CNNs) for environmental sound classification, particularly when dealing with limited amounts of training data, as CNNs have demonstrated superior performance compared to traditional methods and achieve results comparable to other state-of-the-art approaches. (McFee et al. 2014)

  • Exploit the redundancy that exists between different feature channels and filters in convolutional neural networks (CNNs) to improve your efficiency and effectiveness. (Denton et al. 2014)

  • Utilize fully convolutional neural networks (FCNs) for semantic segmentation tasks, as they provide efficient and accurate solutions compared to traditional methods. (Eigen, Puhrsch, and Fergus 2014)

  • Exploit the redundancy that exists between different feature channels and filters in convolutional neural networks (CNNs) to achieve faster computations without compromising accuracy. (Jaderberg, Vedaldi, and Zisserman 2014)

  • Consider using convolutional neural networks (CNN) trained on top of pre-trained word vectors for sentence-level classification tasks, as these models achieve excellent results on multiple benchmarks with minimal hyperparameter tuning and static vectors, and offer even greater performance when learning task-specific vectors through fine-tuning. (Yoon Kim 2014)

  • Consider employing deep learning networks, specifically stacked autoencoders, for EEG-based emotion recognition tasks, as they offer superior performance compared to traditional machine learning models like SVM, particularly when combined with covariate shift adaptation of principal components to address issues related to overfitting and nonstationarity. (Jirayucharoensak, Pan-Ngum, and Israsena 2014)

  • Consider using Acceleration Networks (AccNets) to automate the process of designing fast algorithms for high-dimensional convolution tasks, rather than relying solely on manual approaches. (Aubry et al. 2014)

  • Utilise a fully convolutional encoder-decoder network for object contour detection, which outperforms previous methods in precision and generalises well to unseen object classes within the same super-categories. (L.-C. Chen et al. 2014)

  • Consider increasing the depth of your Convolutional Neural Networks (ConvNets), while maintaining small receptive fields and incorporating many non-linearities, for improved performance in large-scale image recognition tasks. (Simonyan and Zisserman 2014)

  • Utilize group-wise brain damage techniques to improve the efficiency of convolutional neural networks (ConvNets) by modifying the convolutional kernel tensor in a group-wise fashion, leading to faster computations. (Chetlur et al. 2014)

  • Focus on developing a convolution method for point cloud processing that effectively separates the estimation of geometry-less kernel weights and your alignment to the spatial support of features, while also utilizing an efficient point sampling strategy for improved accuracy and computational efficiency. (B. Graham 2014)

  • Utilise Caffe, a flexible, open-source framework that enables efficient and scalable deep learning, facilitated by its modular structure, separation of model representation from implementation, extensive test coverage, and provision of pre-trained reference models. (Yangqing Jia et al. 2014)

  • Consider implementing flattened convolutional neural networks, which involve breaking down traditional 3D convolution filters into three consecutive 1D filters, to achieve faster feed-forward execution without compromising accuracy. (J. Jin, Dundar, and Culurciello 2014)

  • Consider utilizing a Dynamic Convolutional Neural Network (DCNN) for accurate semantic modeling of sentences, as it effectively handles input sentences of varying length, induces a feature graph over the sentence that captures short and long-range relations, and performs well in various language understanding tasks. (Kalchbrenner, Grefenstette, and Blunsom 2014)

  • Employ a dual channel graph convolutional network (DC-GCN) to simultaneously capture both the visual relationships between objects within an image and the syntactic dependencies between words within a question. This approach enables the reduction of semantic gaps between vision and language, leading to improved accuracy in visual question answering tasks. (Diederik P. Kingma and Ba 2014)

  • Carefully examine the necessity of various components within your convolutional neural networks (CNNs), particularly focusing on the potential redundancy of max-pooling layers, which can often be effectively replaced by convolutional layers with increased stride without compromising accuracy across numerous image recognition benchmarks. (Springenberg et al. 2014)

  • Exploit the redundancy that exists between different feature channels and filters in CNNs to achieve faster computations without compromising accuracy. (L. Neumann and Matas 2013)

  • Adopt a two-stage optimization strategy to progressively find good local minima when optimizing a low-precision network, rather than optimizing all aspects simultaneously. (Yoshua Bengio, Léonard, and Courville 2013)

  • Explore alternative methods for constructing deep neural networks on graphs beyond traditional convolutional neural networks, specifically considering spatial and spectral constructions that leverage the unique characteristics of graph-based data. (Bruna et al. 2013)

  • Consider using PointNet, a novel deep learning architecture that directly consumes point clouds, rather than converting them to regular 3D voxel grids or collections of images, as it maintains the permutation invariance of points in the input and offers improved efficiency and effectiveness across a range of 3D classification and segmentation tasks. (Bruna et al. 2013)

  • Consider replacing the traditional fully connected layers in convolutional neural networks with global average pooling layers, as this approach is more native to the convolution structure, avoids overfitting, and is more robust to spatial translations of the input. (M. Lin, Chen, and Yan 2013)

  • Utilise vector quantisation with self-attention for quality-independent representation learning in order to improve the robustness of your deep neural networks against common corruptions. (Yoshua Bengio, Léonard, and Courville 2013)

  • Utilize gradient-based visualization techniques to gain insights into the inner workings of deep convolutional neural networks (ConvNets), enabling them to generate representative images for a class of interest and compute image-specific class saliency maps for weakly supervised object segmentation. (Simonyan, Vedaldi, and Zisserman 2013)

  • Consider implementing deep convolutional neural networks (DNNs) on graphics processing units (GPUs) for efficient and effective image classification tasks, as these networks can significantly outperform traditional methods while requiring less training time. (D. Cireşan, Meier, and Schmidhuber 2012)

  • Consider utilizing convolutional neural networks (ConvNets) with multi-stage features and Lp pooling for image classification tasks, as they offer significant improvements in accuracy when compared to traditional methods. (Sermanet, Chintala, and LeCun 2012)

  • Consider scaling up the core components involved in training deep networks - including the dataset, the model, and the computational resources - in order to effectively learn high-level features from unlabelled data. (Quoc V. Le et al. 2011)

  • Utilize a combination of advanced experimental tools like calcium-sensitive fluorescent indicators and cutting-edge microscopy technologies to observe the simultaneous activity of a large population of neurons, enabling the inference of micro-circuits through the application of efficient computational and statistical methods. (Mishchenko, Vogelstein, and Paninski 2011)

  • Consider utilizing large-scale unsupervised learning techniques to effectively extract high-level features from unlabelled data, leading to improved performance in tasks such as object recognition. (Karo Gregor and LeCun 2010)

  • Utilize Convolutional Networks (ConvNets) for automatic feature learning in order to improve the performance of your machine learning models, especially in areas such as visual perception, auditory perception, and language understanding. (LeCun, Kavukcuoglu, and Farabet 2010)

  • Combine neural architecture search with pruning in a unified approach, known as Sparse Architecture Search (SpArSe), to learn superior models on four popular IoT datasets, resulting in CNNs that are more accurate and up to 4.35 times smaller than previous approaches, while meeting the strict MCU working memory constraint. (Atzori, Iera, and Morabito 2010)

  • Utilise a combination of efficient direct sparse convolution designs, performance modelling, and guided pruning techniques to effectively balance accuracy, speed, and size in convolutional neural networks. (S. Williams, Waterman, and Patterson 2009)

  • Distinguish the contributions of architectures from those of learning systems by reporting random weight performance, as a substantial component of a systems performance can come from the intrinsic properties of the architecture, and not from the learning system.’ (Gray 2005)

  • Consider utilizing deep neural networks (DNNs) for weather forecasting tasks, particularly those involving precipitation, due to your ability to handle large spatial and temporal contexts, provide probabilistic outputs representing uncertainty, and adapt easily to increasing amounts of training data. (“A Vision for the National Weather Service” 1999)

  • Adopt a Bayesian approach to modeling and classifying neural signals, allowing them to infer a probabilistic model of the waveform, quantify the uncertainty of the form and number of inferred action potential shapes, and efficiently decompose complex overlaps. (Lewicki 1994)

  • Consider using a Hierarchical Gaussian Mixture representation for adaptive 3D registration tasks, as it allows for efficient and accurate point cloud data processing across a range of complex environments. (Besl and McKay 1992)

  • Focus on visualizing invariance in deep neural networks alongside selectivity, as it offers valuable insights into the computations performed by these systems. (Adelson and Bergen 1985)

  • Consider using a bilateral neural network (Bi-NN) framework for cross-language algorithm classification, which involves building a neural network on top of two underlying sub-networks, each encoding syntax and semantics of code in one language, and training the whole Bi-NN with bilateral programs that implement the same algorithms and/or data structures in different languages. (K. L. Clark 1980)

  • Consider the potential influence of adaptivity and distribution gaps when interpreting the generalizability of machine learning models based on test set performance. (NA?)

  • Explore deep learning architectures instead of shallow ones, as deep architectures have the potential to generalize in non-local ways, allowing for greater scalability and applicability to complex tasks. (NA?)

  • Strive to create machines that learn and think like humans by focusing on three core elements: building causal models of the world, grounding learning in intuitive theories of physics and psychology, and leveraging compositionality and learning-to-learn to rapidly acquire and generalize knowledge to new tasks and situations. (NA?)

  • Consider adopting a deep architecture for matching short texts, which enables explicit capture of natural nonlinearities and hierarchical structures in matching two structured objects. (NA?)

  • Utilize the Thermodynamic Bethe Ansatz (TBA) to analyze the area of minimal surfaces in AdS space, as it provides an effective framework for understanding the relationship between the area and the shape of the polygon. (NA?)

  • Use an integrable spin-chain model to accurately calculate the full function of cusped Wilson loops in the planar approximation, as it provides a comprehensive framework for understanding the behavior of these systems. (NA?)

  • Consider using deep learning techniques, specifically deep belief networks (DBNs) and convolutional neural networks (CNNs), to efficiently handle and analyze massive amounts of data, taking advantage of the increased processing power provided by graphics processors and other high-performance computing resources. (NA?)

  • Consider using deep dynamic neural networks (DDNN) for multimodal gesture recognition, which involves a semi-supervised hierarchical dynamic framework based on a Hidden Markov Model (HMM) for simultaneous gesture segmentation and recognition, leveraging skeleton joint information, depth and RGB images as multimodal input observations. (NA?)

  • Adopt a combination of competitive and cooperative mechanisms within a crowdsourcing framework to effectively develop and refine advanced algorithms for analyzing complex neuroimaging data. (NA?)

  • Consider utilizing convolutional neural networks for the classification of electromyography data, as they demonstrate superior performance compared to traditional classification methods in the context of prosthetic hand control. (NA?)

  • Consider implementing probabilistic weighted pooling instead of max-pooling in convolutional neural networks, as it leads to improved accuracy through efficient model averaging. (NA?)

  • Utilize deep learning techniques, specifically deep neural networks, to improve the accuracy of predicting DNA methylation states from DNA sequence and incomplete methylation profiles in single cells. (NA?)

  • Employ a unified discriminative framework using a deep convolutional neural network to classify gene expression using histone modification data as input, allowing for the simultaneous visualization of combinatorial interactions among histone modifications. (NA?)

  • Adopt the “Learning without Forgetting” (LwF) method when they need to add new capabilities to a Convolutional Neural Network (CNN) without losing the original capabilities, even when the training data for those original capabilities is unavailable. (NA?)

  • Consider adopting a Bayesian probabilistic perspective when working with deep learning models, as it offers several advantages including improved efficiency in algorithm optimization and hyper-parameter tuning, as well as enhanced predictive performance through the utilization of multiple deep layers of data reduction. (NA?)

  • Consider using deep learning techniques to improve the accuracy of protein function prediction, especially when dealing with large-scale, multi-class, multi-label problems like those encountered in the Gene Ontology. (NA?)

  • Focus on developing image processing-based plant disease identification systems that can diagnose diseases in your early development stages, increasing the reliability of disease identification and validating it on real environments. (NA?)

  • Carefully consider the choice of deep ConvNet architecture, incorporating recent advancements like batch normalization and exponential linear units, along with a cropped training strategy, to achieve optimal decoding performance for EEG analysis. (NA?)

  • Focus on developing end-to-end multimodal emotion recognition systems using deep neural networks, specifically incorporating auditory and visual modalities, to achieve superior performance in accurately identifying emotional states. (NA?)

  • Focus on developing models that can automatically learn features for sleep stage scoring from different raw single-channel EEGs from various datasets without requiring any hand-engineered features. (NA?)

  • Utilise the newly developed Semantic3D.net’, a large-scale point cloud classification benchmark data set containing over four billion manually labelled points, as input for data-hungry deep learning methods to enhance your performance in 3D point cloud labelling tasks.’ (NA?)

  • Utilise a novel visualisation framework to create groups of clusters or summaries’, each containing crisp salient image regions that focus on a particular aspect of an image class that the network has exploited with high regularity. This enables clearer communication about what a network has learned about a particular image class, and can help improve classification accuracy.’ (“A 5b 800MS/s 2mW Asynchronous Binary-Search ADC in 65nm CMOS,” n.d.)

  • Carefully consider the trade-off between efficiency and rotation equivariance when designing convolutions for spherical neural networks, and that using a graph-based spherical CNN like DeepSphere provides a flexible and effective balance between these two factors. (NA?)

  • Carefully choose and optimize the hyperparameters of your Convolutional Neural Networks (CNNs) for sentence classification tasks, as significant variations in performance can occur depending on the chosen configuration. (NA?)

  • Focus on developing novel representations of filters, like Filter Summary (FS), that enforce weight sharing across filters to achieve model compression while maintaining high performance in deep Convolutional Neural Networks (CNNs). (NA?)

  • Utilise a combination of deep learning algorithms, specifically Convolutional Neural Networks (CNNs), along with linguistic patterns to achieve superior results in aspect extraction tasks compared to traditional methods. (NA?)

  • Utilise deep neural networks (DNNs) to improve the accuracy of click-through rate (CTR) predictions in online display advertising. (NA?)

  • Consider using a mini-batch aware regularizer to save heavy computation of regularization on deep networks with huge numbers of parameters, while also employing a data adaptive activation function to generalize PReLU by considering the distribution of inputs, ultimately leading to improved performance in training industrial deep networks. (NA?)

  • Consider using a modular decoding approach, which involves constructing multi-scale local decoders that predict the contrast of local image patches, to enable the reconstruction of arbitrary visual images from brain activity. (NA?)

  • Consider using neural networks to identify and differentiate various phases of matter, including both conventional ordered phases and unconventional phases like those found in the square-ice model and the Ising lattice gauge theory, due to your ability to learn the order parameters of these phases without explicit knowledge of the energy or locality conditions of the Hamiltonian. (NA?)

  • Employ layer-wise relevance propagation (LRP) to trace the classification decision back to individual words in text documents, enabling a deeper understanding of the categorization process and facilitating the generation of novel vector-based document representations that capture semantic information. (NA?)

  • Carefully choose appropriate compression and decompression techniques for reducing the dimensionality of label vectors in extreme multi-label text classification (XMTC) tasks, as this can greatly impact the efficiency and reliability of learned mappings from feature space to compressed label space. (NA?)

  • Incorporate Bayesian model uncertainty into your analysis, as it provides valuable additional information beyond traditional network outputs, allowing for improved decision making and increased accuracy in predictions. (NA?)

  • Consider applying deep convolutional neural networks (DCNNs) for Raman spectrum recognition, as they provide a unified solution that eliminates the need for ad-hoc preprocessing steps and demonstrates superior classification performance compared to other commonly used machine learning algorithms like support vector machines. (NA?)

  • Consider using a fusion convolutional long short-term memory network (FCL-Net) for short-term passenger demand forecasting in on-demand ride services, as it effectively captures spatio-temporal characteristics and correlations of explanatory variables, leading to improved predictive performance. (NA?)

  • Carefully choose the appropriate cost function for your specific application, considering factors such as cross-entropy loss for classification problems and generative adversarial networks for image prediction tasks, to ensure accurate and reliable results. (NA?)

  • Consider adopting deep learning architectures like Convolutional Neural Networks (CNNs) when attempting to predict drug-target binding affinities, as demonstrated by the superior performance of the proposed DeepDTA model in comparison to traditional machine learning algorithms and other deep learning approaches. (NA?)

  • Consider utilizing advanced machine learning techniques like deep learning (DL), reinforcement learning (RL), and your combination (deep RL) for effectively handling and interpreting complex biological data. (NA?)

  • Utilize an ensemble of deep convolutional neural networks (DCNNs) to enhance the accuracy of skin lesion classification, particularly for melanoma detection. (NA?)

  • Carefully choose an appropriate deep learning architecture for medical image analysis tasks based on the number of available images and ground truth labels. (NA?)

  • Utilize deep learning algorithms for sentiment analysis of financial data, particularly when dealing with large amounts of unstructured data, as it allows for the extraction of complex data at a high level of abstraction and can be invariant to local changes in the input data. (NA?)

  • Consider utilizing deep convolutional neural networks (CNNs) for the automated diagnosis and prediction of periodontitis compromised teeth (PCT) in periapical radiographs, achieving comparable accuracy to experienced periodontists. (NA?)

  • Utilize deep learning algorithms, specifically Convolutional Neural Networks (CNNs), to improve the accuracy and efficiency of galaxy morphological classification, particularly for large datasets like the Sloan Digital Sky Survey. (NA?)

  • Apply oversampling to eliminate class imbalance in convolutional neural networks, while considering the optimal undersampling ratio depending on the degree of imbalance, without causing overfitting. (NA?)

  • Consider using a fully convolutional neural network (FCNN) for direct white matter tract segmentation, as it offers complete and accurate segmentations while being easier to set up, faster to run, and not requiring registration, parcellation, tractography, or clustering. (NA?)

  • Consider utilising deep learning algorithms, specifically convolutional neural networks (CNNs), for efficient and accurate classification of echocardiogram views, potentially improving diagnostics and treatment planning in cardiovascular diseases. (NA?)

  • Use a combination of a novel triplet selection module called “Group Hard” for effective triplet training, a standard deep convolutional neural network for learning deep representations, a well-specified triplet loss for pulling together similar pairs and pushing away dissimilar pairs, and a novel triplet quantization loss with weak orthogonality constraint for converting the deep representations of different samples into B-bit compact binary codes, ultimately leading to state-of-the-art retrieval results on various image (NA?)

  • Consider utilizing deep learning algorithms, particularly convolutional and recurrent neural networks, to analyze medical imagery for improved prognostic stratification and disease subtyping, potentially leading to more accurate and personalized treatments. (NA?)

  • Focus on developing an optical convolutional (opt-conv) layer with an optimizable phase mask that leverages the inherent convolution performed by a linear, spatially invariant imaging system, enabling low-power inference by a custom optoelectronic CNN. (NA?)

  • Consider combining Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) architectures to achieve higher accuracy in predicting particulate matter (PM2.5) concentrations in smart cities. (NA?)

  • Consider utilizing the SchNet deep learning architecture for modeling complex atomic interactions in order to predict potential-energy surfaces or speed up the exploration of chemical space, as it follows fundamental symmetries of atomistic systems by construction and enables accurate predictions throughout compositional and configurational chemical space. (NA?)

  • Focus on developing a scalable solution for both computation and memory architectures on high-end FPGAs, aiming to reduce deployment costs for different models using a general-purpose design. (NA?)

  • Consider employing deep learning techniques, particularly those involving neural networks, convolutional neural networks, recurrent neural networks, long short term memory, gated recurrent units, autoencoders, restricted boltzmann machines, and generative adversarial networks, due to your ability to handle complex data sets, adapt to changing conditions, generalize across various contexts, and scale effectively. (NA?)

  • Pay attention to the unique properties of hyperspectral data, including its high spectral resolution, low spatial resolution, and relatively small data volumes, when developing deep learning models for classification tasks. (NA?)

  • Consider integrating user interactions into CNN frameworks to obtain accurate and robust segmentation of 2D and 3D medical images while making the interactive framework more efficient with a minimal number of user interactions. (NA?)

  • Consider using deep learning techniques, particularly convolutional neural networks (CNNs), for medical image segmentation tasks, as they offer significant improvements in accuracy and efficiency over traditional methods. (NA?)

  • Consider combining classical and learning-based methods in order to achieve accurate, fast, and topology-preserving image registration. (NA?)

  • Consider using deep contextual learning for base-pair prediction, particularly for non-canonical and non-nested (pseudoknot) base pairs stabilized by tertiary interactions, and leverage transfer learning from a model initially trained with a high-quality bpRNA dataset to achieve statistically significant improvements in predicting all types of base pairs. (NA?)

  • Consider utilizing deep learning techniques, specifically convolutional neural networks and long short-term memory networks, to improve the accuracy and speed of protein structural feature predictions. (NA?)

  • Consider utilizing a deep residual network of convolutional and recurrent units for earthquake signal detection, as it enables automatic extraction of sparse features from seismograms, provides robust models for sequential characteristics of seismic data, prevents degradation, reaches higher accuracy with deeper learning, and demonstrates superior performance in the presence of high noise levels compared to other traditional methods. (NA?)

  • Consider utilizing larger, more diverse datasets, employing data augmentation techniques such as GANs, and focusing on improving the accuracy of plant disease detection in real-world environments through innovative neural network architectures. (NA?)

  • Consider utilizing the DeTraC deep convolutional neural network for accurate classification of COVID-19 in chest X-ray images, particularly when dealing with data irregularities. (NA?)

  • Combine the strengths of LSTM and CNN models with an added attention mechanism to achieve greater accuracy in text classification tasks. (NA?)

  • Consider using multi-objective differential evolution (MODE) to optimize the hyperparameters of convolutional neural networks (CNN) for accurate classification of COVID-19 patients from chest CT images. (NA?)

  • Use appropriate validation techniques like k-fold cross-validation to prevent overfitting of data, and consider utilizing AI-based algorithms to enhance diagnostic accuracy and potentially improve patient outcomes in areas such as gastroenterology and hepatology. (NA?)

  • Carefully select appropriate deep learning algorithms for specific application scenarios, considering factors like the setup environment, data size, and number of sensors and sensor types, to optimize the performance of bearing fault diagnostic systems. (NA?)

  • Use Topaz-Denoise, a deep learning method based on a pre-trained general model, to effectively denoise cryoEM images and cryoET tomograms, thereby improving micrograph interpretability, enabling faster data collection, and facilitating downstream analysis. (NA?)

  • Use transfer learning when working with limited data sets in order to improve the accuracy of your predictions. (NA?)

  • Utilise a combination of convolutional neural networks (CNNs) and progressive generative adversarial networks (GANs) to effectively analyse and manipulate image data, enabling accurate classification and manipulation of visual elements. (NA?)

  • Consider employing transfer learning with convolutional neural networks, particularly those well-trained on non-medical ImageNet datasets, when working with medical image analysis tasks where large labeled datasets are unavailable or insufficient. (NA?)

  • Carefully consider the appropriate selection of machine learning algorithms and deep learning architectures based on the specific problem, data type, and desired outcome, taking into account factors such as performance, computational resources, and interpretability. (NA?)

  • Carefully consider the choice of convolutional neural network (CNN) architecture, taking into account factors such as spatial exploitation, depth, multiple paths, feature-map exploitation, width, attention mechanisms, and dimension-based optimization, in order to achieve optimal performance in computer vision tasks. (NA?)

  • Consider applying convolutional neural networks (CNNs) in your medical image understanding studies, as they have demonstrated superior performance in numerous applications, including image classification, segmentation, localization, and detection, and have the potential to improve diagnoses and reduce medical trauma. (NA?)

  • Carefully consider the location of task interactions in your multi-task learning architectures, distinguishing between encoder-focused and decoder-focused models, to optimize performance in dense prediction tasks. (NA?)

  • Utilize deep learning techniques, particularly convolutional neural networks, to effectively analyze and interpret vast amounts of data in various fields, thereby improving overall model performance and reducing reliance on manual feature engineering. (NA?)

  • Consider utilizing federated learning (FL) for developing artificial intelligence (AI) models in healthcare settings, particularly for cross-institutional studies, as it allows for efficient data collaboration without compromising data privacy and security. (NA?)

  • Consider using a combination of guilt-by-association heuristics and machine-learning techniques to effectively detect and characterize scam tokens within decentralized exchanges. (NA?)

  • Consider using machine learning techniques, particularly transfer learning, to efficiently and effectively detect vulnerabilities in smart contracts, allowing for faster adaptation to new vulnerability types with limited data. (NA?)

  • Utilize deep convolutional neural networks (DCNNs) to separate and recombine the image content and style of natural images, allowing them to produce new images of high perceptual quality that combine the content of an arbitrary photograph with the appearance of numerous well-known artworks. (NA?)

Recurrent Neural Networks (Rnn)

  • Consider using domain-specific word embeddings along with a bidirectional LSTM-based deep model as a classifier for automatic detection of hate speech, achieving a 93% F1-score, while also evaluating the effectiveness of transfer learning language model (BERT) on the hate speech problem as a binary classification task, achieving a 96% F1-score on a combined balanced dataset from available hate speech datasets. (H. Saleh, Alhothali, and Moria 2023)

  • Consider using prompt engineering-assisted malware dynamic analysis with GPT-4 to generate explanatory text for each API call within the API sequence, followed by applying BERT to obtain the representation of the text, and finally using a CNN-based detection model to extract the feature. (P. Yan et al. 2023)

  • Consider the impact of frame-level changes on token-level sequences when estimating uncertainty in connectionist temporal classification (CTC)-based automatic speech recognition models, as this leads to improved accuracy in recognizing errors. (Rumberg et al. 2023)

  • Use Merlion, an open-source machine learning library for time series, which offers a unified interface for various models and datasets, standard pre/post-processing layers, visualization, anomaly score calibration, AutoML for hyperparameter tuning and model selection, and model ensembling, allowing for rapid development and benchmarking of models across multiple time series datasets. (Bhatnagar et al. 2021)

  • Consider combining graph convolutional networks (GCNs) and recurrent neural networks (RNNs) to model the information diffusion process of article links in order to achieve improved results in tasks such as rumor detection. (D. Huang, Bartel, and Palowitch 2021)

  • Utilise a combination of recurrent and graph neural network architectures to jointly model time and graph information in dynamic graph data, whilst employing a scalable training scheme and self-supervised pretraining framework to enhance model performance and address issues of label scarcity. (A. Z. Wang et al. 2021)

  • Incorporate a deep spatio-temporal and contextual neural network called DeepFEC to accurately predict energy consumption in transportation networks, accounting for various factors such as vehicle type, road topology, traffic, vehicle speed, driving style, ambient temperature, road conditions, and road grade. (Elmi and Tan 2021)

  • Consider utilizing deep learning techniques, particularly neural networks, for time series forecasting due to your ability to effectively capture complex patterns and relationships within the data. (Theodosiou and Kourentzes 2021)

  • Consider using domain-wall memory (DWM) for efficient acceleration of recurrent neural networks (RNNs), as it offers high density, linear access patterns, and low read/write energy. (Samavatian et al. 2020)

  • Utilize a Total Probability Formula and Adaptive GRU Loss Function based Deep Neural Network (TPG-DNN) for user intent prediction. (J. Jiang et al. 2020)

  • Modify the RNN-T loss function to develop Alignment Restricted RNN-T (Ar-RNN-T) models, which utilize audio-text alignment information to guide the loss computation, improving downstream applications such as the ASR End-pointing by guaranteeing token emissions within any given range of latency. (Mahadeokar et al. 2020)

  • Utilize the Long Short-Term Memory (LSTM) network instead of the Gated Recurrent Unit (GRU) for the task of algorithmic music generation, as the former produces significantly more musically plausible outputs. (Gunawan, Iman, and Suhartono 2020)

  • Consider using hierarchical recurrent neural networks (HRNNs) for efficient and accurate modelling of time series data, especially in cases involving large item catalogues and cold-start scenarios. (Yifei Ma et al. 2020)

  • Consider implementing a modular architecture, such as MASR, when working with sparse RNNs for automatic speech recognition tasks. (U. Gupta et al. 2019)

  • Utilize the concept of an “action graph” to model user behaviour in mobile social apps, as it allows for a more comprehensive understanding of user engagement patterns than traditional macroscopic approaches. (Yozen Liu et al. 2019)

  • Utilize the KBLSTM model, which combines bi-directional LSTMs with an attention mechanism and a sentinel component, to effectively integrate background knowledge from external knowledge bases into machine reading tasks, thereby improving overall performance. (Bishan Yang and Mitchell 2019)

  • Utilize a novel approach called “JODIE” (Joint Dynamic User-Item Embeddings) to improve the accuracy and efficiency of recommendation systems. This involves using a coupled recurrent neural network model to learn embedding trajectories of users and items, along with a projection operator to predict future interactions in constant time. Additionally, the authors suggest implementing a batching algorithm called “t-Batch” to speed up the training process by creating independent but temporally consistent training data batch (S. Kumar, Zhang, and Leskovec 2019)

  • Consider using a combination of the User Interest Center (UIC) module and the Multi-channel user Interest Memory Network (MIMN) architecture to effectively handle long sequential user behavior data for click-through rate (CTR) prediction tasks. (Pi et al. 2019)

  • Consider utilising a stochastic recurrent neural network for multivariate time series anomaly detection, specifically the OmniAnomaly model, which effectively deals with explicit temporal dependence among stochastic variables to learn robust representations of input data. (Ya Su et al. 2019)

  • Consider using a shallow gated recurrent unit (GRU) neural network architecture for eating detection tasks on low power micro-controllers, as it provides high accuracy while conserving memory and computational resources. (Amoh and Odame 2019)

  • Consider utilizing a combination of multilevel discrete wavelet decomposition (MDWD) and deep learning techniques, specifically recurrent neural networks (RNN) and long short-term memory (LSTM), to effectively analyze complex time series data. (Jingyuan Wang et al. 2018)

  • Focus on developing a comprehensive framework like MSCRED that addresses multiple aspects of anomaly detection and diagnosis simultaneously, including temporal dependency, noise resistance, and severity interpretation, rather than tackling each aspect separately. (Tianyun Zhang, Ye, Zhang, Tang, et al. 2018)

  • Develop a global optimization framework for mutual influence aware ranking in e-commerce search, focusing on directly optimizing the Gross Merchandise Volume (GMV) for ranking and decomposing ranking into two tasks: mutual influence aware purchase probability estimation and finding the best ranking order based on the purchase probability estimations. (T. Zhuang, Ou, and Wang 2018)

  • Utilise a Long Short Term Memory (LSTM) model for electric load forecasting, enhanced by feature selection and genetic algorithm (GA) to optimize time lags and number of layers, resulting in increased forecasting accuracy. (Bouktif et al. 2018)

  • Utilise deep bidirectional recurrent neural networks (DBRNN) and deep bidirectional long short term memory (DBLSTM) architectures for speaker-adapted confidence measures in automatic speech recognition (ASR) systems. This is due to your ability to efficiently model temporal dependencies, handle vanishing gradient problems, and incorporate context information in both time directions, leading to significant improvements in classification error rates and normalised cross entropy scores. (Del-Agua et al. 2018)

  • Utilize a rule-based method to predict candidate arguments on the event types of possibilities, followed by application of a recurrent neural network model called RNN-ARG with the attention mechanism for event detection to effectively capture meaningful semantic regularities from these predicted candidate arguments. (Wentao Wu et al. 2018)

  • Consider using block-circulant matrices for structured compression of LSTM models, enabling faster computation and reduced memory usage without compromising accuracy. (Shuo Wang et al. 2018)

  • Consider using deep neural networks to automatically infer the syntax and semantics of programming languages from large corpora of human-generated code, rather than relying on laborious and potentially incomplete expert-defined grammars. (Cummins et al. 2018)

  • Leverage natural language information in source code, such as comments, function names, and parameter names, to enhance type inference accuracy in dynamically typed languages like JavaScript. (Ore et al. 2018)

  • Utilize a multi-level attention-based recurrent neural network when attempting to predict geo-sensory time series, as it effectively accounts for dynamic spatio-temporal correlations and external factors. (Yuxuan Liang et al. 2018)

  • Consider using deep neural networks, specifically recurrent neural networks (RNNs), for making continual predictions based on raw mobile phone sensor data, as demonstrated by the success of this approach in accurately predicting notification attendance. (Katevas et al. 2017)

  • Consider using a generative model with an encoder-decoder framework for keyphrase prediction, as it can effectively overcome the limitations of traditional approaches by identifying keyphrases that do not appear in the text and capturing the true semantic meaning behind the text. (R. Meng et al. 2017)

  • Utilize the weight-dropped LSTM, which employs DropConnect on hidden-to-hidden recurrent weights, along with NT-ASGD, a variant of the averaged stochastic gradient method, to optimize and regularize LSTM-based models for improved word-level language modeling performance. (Merity, Keskar, and Socher 2017)

  • Utilize a novel decomposition of the output of an LSTM into a product of factors, assigning importance scores to words according to your contribution to the LSTMs prediction, enabling the identification of consistently important patterns of words, and ultimately leading to the creation of a simple, rule-based classifier that closely approximates the output of the LSTM.’ (Murdoch and Szlam 2017)

  • Focus on identifying and mitigating sources of bias in production speech models through improved neural architectures for streaming inference, optimisation techniques, and increased audio and label modelling versatility. (Battenberg et al. 2017)

  • Incorporate recursion into your neural architectures to enhance generalizability and interpretability, particularly when dealing with complex input structures. (J. Cai, Shin, and Song 2017)

  • Consider utilizing a long short-term memory-based variational autoencoder (LSTM-VAE) for multimodal anomaly detection, as it effectively combines both temporal and spatial information, allowing for more accurate identification of anomalies within complex datasets. (D. Park, Hoshi, and Kemp 2017)

  • Consider using a multi-scale language model that combines global and local features to improve extraction of key information from ontologies, leading to greater processing efficiency and higher performance than traditional single RNN layer models. (Yukun Yan et al. 2017)

  • Consider using heterogeneous information network (HIN)-compatible recurrent neural networks (RNNs) for fraudster group detection, as it allows for the encoding of non-local semantic dependencies between reviewers through an autoregressive model, leading to improved accuracy in identifying fraudulent groups. (Yafeng Ren and Ji 2017)

  • Consider developing mobile-specific optimization frameworks for recurrent neural network (RNN) models, such as MobiRNN, to efficiently execute them on mobile GPUs, taking into account factors like device type, model complexity, and GPU load. (Q. Cao, Balasubramanian, and Balasubramanian 2017)

  • Develop a dynamic, hierarchically scoped, open vocabulary language model for source code, utilizing mixed, scoped models and a fast data structure optimized for dynamic, scoped counting of language events, to achieve best-in-class performance in non-parametric (count-based) language modeling. (Hellendoorn and Devanbu 2017)

  • Employ a combination of log key anomaly detection and parameter value anomaly detection models, along with a workflow model, to effectively identify and diagnose anomalies in system logs. (Min Du et al. 2017)

  • Utilise a LSTM-based model for sentiment analysis in videos, allowing utterances to capture contextual information from your surroundings in the same video, thereby significantly improving the classification process. (Poria et al. 2017)

  • Consider utilizing deep learning models to analyze large datasets of historical peer reviews in order to develop intelligent code review systems capable of identifying and recommending relevant reviews for specific code snippets, thereby improving the efficiency and effectiveness of the code review process. (Allamanis et al. 2017)

  • Utilise the recurrent relational network (RRN) model for tasks involving multiple steps of relational reasoning, as demonstrated by its success in solving complex tasks such as Sudoku puzzles and answering complex questions about relationships between objects. (B. Amos and Kolter 2017)

  • Consider using residual LSTM architecture for deep recurrent neural networks, as it offers an additional spatial shortcut path for efficient training while reducing network parameters by more than 10%. (Jaeyoung Kim, El-Khamy, and Lee 2017)

  • Consider using the Long- and Short-term Time-series Network (LSTNet) for multivariate time series forecasting, as it effectively combines the strengths of convolutional layers for local dependency discovery and recurrent layers for complex long-term dependency capture, while also incorporating a traditional autoregressive linear model for increased robustness against scale changes. (Lai et al. 2017)

  • Utilize a dual-stage attention-based recurrent neural network (DA-RNN) for effective time series prediction, as it allows for adaptive extraction of relevant driving series and selection of relevant encoder hidden states across all time steps. (Yao Qin et al. 2017)

  • Carefully consider the impact of distributional issues and limited model capacities when comparing the performance of unsupervised versus supervised approaches in representation learning, particularly for tasks involving sentiment analysis. (Radford, Jozefowicz, and Sutskever 2017)

  • Utilize auxiliary prediction tasks to evaluate and compare various sentence embedding techniques, focusing on fundamental sentence properties like length, word content, and word order. (Adi et al. 2016)

  • Combine symbolic knowledge provided by knowledge graphs with RNN language models to improve the perplexity and reduce the number of unknown words in language modeling. (S. Ahn et al. 2016)

  • Consider utilizing the Dynamic Memory Network (DMN) architecture when working on natural language processing tasks, as it enables the processing of input sequences and questions, formation of episodic memories, and generation of relevant answers through an iterative attention process and hierarchical recurrent sequence model. (Andreas et al. 2016)

  • Develop deep learning models like GRU-D to effectively exploit two representations of informative missingness patterns, i.e., masking and time interval, in order to improve prediction results in time series classification tasks. (Z. Che et al. 2016)

  • Consider using a combination of multi-relational embedding-based models, such as TransE, and recurrent neural networks with attention mechanisms to generate high-quality factoid question-answer pairs for training question-answering systems. (Serban, García-Durán, et al. 2016)

  • Consider using recurrent neural networks (RNNs) instead of traditional vector-based methods for analyzing consumer behavior in e-commerce, because RNNs can handle variable-length sequences, reduce the need for manual feature engineering, and provide greater interpretability of predictions. (Wangperawong et al. 2016)

  • Consider using Recurrent Neural Networks (RNNs) dedicated to individual attributes rather than concatenating attribute word sequences, as this approach improves the ability of the model to capture the full meaning of text descriptions and reduces ambiguity. (J.-W. Ha, Pyo, and Kim 2016)

  • Utilise the Professor Forcing algorithm when training recurrent networks, as it encourages the dynamics of the network to remain consistent during training and sampling, thereby acting as a regulariser and improving overall performance. (Lamb et al. 2016)

  • Consider using a recurrent neural network architecture like RaSoR to build efficient, fixed-length span representations of all possible answer spans within a given evidence document, which can lead to improved performance in tasks involving answer extraction from text. (Kenton Lee et al. 2016)

  • Frame the few-shot learning problem within a meta-learning setting, utilizing an LSTM-based meta-learner optimizer to optimize a learner neural network classifier, thereby addressing the limitations of traditional gradient-based optimization approaches. (Oord, Dieleman, et al. 2016)

  • Extend the sequence-to-sequence framework to model natural language generation as two parallel discrete stochastic processes: a sequence of high-level coarse tokens, and a sequence of natural language tokens. (Serban, Klinger, et al. 2016)

  • Optimize your models using both supervised learning and reinforcement learning techniques, as they are complementary and can significantly enhance the learning rate and overall performance of the model. (J. D. Williams and Zweig 2016)

  • Utilise a sequence-to-sequence model for user simulation in spoken dialogue systems, as it effectively addresses limitations of previous models by taking into account the entire dialogue history, ensuring coherent user behavior without reliance on external data structures, and allowing for modelling of user behavior with finer granularity. (Asri, He, and Suleman 2016)

  • Utilize a deep learning model, specifically a Sequence-to-Sequence model, to automatically generate syntactically valid C programs for fuzz testing, thereby increasing the efficiency and effectiveness of compiler testing. (Sahil Bhatia and Singh 2016)

  • Focus on developing end-to-end dialog systems capable of handling goal-oriented dialogues, specifically in the context of restaurant reservations, through the use of Memory Networks, which have demonstrated promising performance in non-goal oriented dialogue. (Bordes, Boureau, and Weston 2016)

  • Use a time-adaptive recurrent neural network (TARN) to learn to modulate time constants of transition function, allowing for selectively pondering on informative inputs to strengthen your contribution, and ignoring noisy inputs. This modification, along with designing suitable transition matrices, yields lossless information propagation, improving trainability and handling of long-term dependency tasks with a lighter memory footprint. (Bradbury et al. 2016)

  • Consider using Long Short-Term Memory-Networks (LSTMNs) for machine reading tasks, as they enable adaptive memory usage during recurrence with neural attention, thereby improving the understanding of structured input. (Jianpeng Cheng, Dong, and Lapata 2016)

  • Employ a fully probabilistic treatment of the problem with a novel conditional parameterization using neural networks, propose the focused pruning method to reduce the search space during inference, and investigate two variations to improve the generalization of representations for millions of entities under highly sparse supervision. (Z. Dai, Li, and Xu 2016)

  • Consider utilizing a novel deep learning model that captures the nonlinear coevolution nature of users and items’ embeddings in a nonparametric manner, assigning an evolving feature embedding process for each user and item, and modeling the co-evolution of these latent feature processes with two parallel components: (i) item → user component, where a user’s latent feature is determined by the nonlinear embedding of latent features of the items he interacted (H. Dai et al. 2016)

  • Consider utilizing recurrent neural network grammars (RNNGs) for improved parsing and language modeling performance, as demonstrated by your superior results compared to other existing methods. (Dyer et al. 2016)

  • Consider using a hierarchical encoder-decoder model to improve the quality of sentence representations by capturing longer-term dependencies between sentences. (Gan et al. 2016)

  • Incorporate the concept of “Adaptive Computation Time” (ACT) into your recurrent neural network models, enabling these models to learn the optimal number of computational steps to take between receiving an input and producing an output, thereby improving overall performance. (Graves 2016)

  • Employ a deep learning approach called DeepAPI, which leverages a neural language model called RNN Encoder-Decoder, to accurately generate API usage sequences for a given natural language query. (X. Gu et al. 2016)

  • Focus on developing effective quantization methods for recurrent neural networks (RNNs) to reduce bit-widths of weights, activations, and gradients, thereby improving storage size, memory usage, and training/inference speeds while maintaining or even enhancing overall performance. (Qinyao He et al. 2016)

  • Explore the potential of bilinear LSTM models for improving the learning of long-term appearance models in multi-object tracking applications, as it allows for a multiplicative coupling between the memory and the input, mimicking an online learned classifier/regressor at each time step. (Keuper et al. 2016)

  • Consider using an LSTM-based Encoder-Decoder scheme for Anomaly Detection (EncDec-AD) in multi-sensor time-series, as it effectively learns to reconstruct normal’ time-series behavior and uses reconstruction error to identify anomalies, proving to be robust across various types of time-series including those that are predictable, unpredictable, periodic, aperiodic, and quasi-periodic.’ (Malhotra et al. 2016)

  • Consider using a hierarchical framework of memory-less autoregressive multilayer perceptrons and stateful recurrent neural networks to effectively capture underlying sources of variation in temporal sequences across various datasets. (Mehri et al. 2016)

  • Utilise Pixel Recurrent Neural Networks (PixelRNNs) when modelling the distribution of natural images due to your ability to sequentially predict pixels in an image along the two spatial dimensions while encoding the complete set of dependencies in the image. (Oord, Kalchbrenner, and Kavukcuoglu 2016)

  • Utilize the Query-Reduction Network (QRN) approach for question answering tasks requiring reasoning over multiple facts, as it effectively manages both short-term and long-term sequential dependencies, outperforms existing methods, and offers potential for parallelization. (Seo, Min, et al. 2016)

  • Consider using end-to-end attention-based models with multichannel input and Highway long short-term memory (HLSTM) for improved performance in Distant Speech Recognition tasks. (Taherian 2016)

  • Focus on developing strong patch-based residual encoders and entropy coders capable of capturing long-term dependencies between patches in the image to improve compression rates for a given quality. (Toderici et al. 2016)

  • Consider using deep spatio-temporal residual networks (ST-ResNet) to collectively predict inflow and outflow of crowds in every region of a city, taking into account spatial dependencies, temporal dependencies, and external influences. (Junbo Zhang, Zheng, and Qi 2016)

  • Consider utilizing character-based word embeddings in your models, as opposed to traditional word embeddings, to better capture the morphology of words in morphologically rich languages. (Ballesteros, Dyer, and Smith 2015)

  • Consider using recurrent neural networks (RNNs) for predicting diagnoses, medications, and visit times in electronic health records (EHRs), as demonstrated by the authors achieving promising results in your study. (E. Choi et al. 2015)

  • Consider employing a character-aware neural language model when working with languages that have rich morphologies, as it can lead to improved performance while requiring fewer parameters compared to other approaches. (Yoon Kim et al. 2015)

  • Consider utilizing the Dynamic Memory Network (DMN) architecture when working on natural language processing tasks, as it enables the processing of input sequences and questions, formation of episodic memories, and generation of relevant answers through an iterative attention process and hierarchical recurrent sequence model. (A. Kumar et al. 2015)

  • Carefully balance the competing goals of learning and fuzzing in your experimental designs, recognizing that learning seeks to capture the structure of well-formed inputs, while fuzzing aims to break that structure in order to identify unexpected code paths and potential bugs. (Kurach, Andrychowicz, and Sutskever 2015)

  • Utilize Long Short-Term Memory (LSTM) recurrent neural networks for analyzing multivariate time series of clinical measurements, as they effectively model varying length sequences and capture long range dependencies, leading to improved performance compared to traditional methods. (Lipton et al. 2015)

  • Consider extending LSTM to tree structures when dealing with complex input structures, as doing so allows for the reflection of historical memories of multiple child and descendant cells, leading to improved performance in tasks such as semantic composition. (X. Zhu, Sobhani, and Guo 2015)

  • Consider utilizing stack LSTMs, a novel extension of traditional LSTMs, to enhance the representational capacity of your models. By incorporating a stack pointer mechanism, stack LSTMs allow for greater flexibility in processing sequential data, enabling improved performance across various natural language processing tasks. (Dyer et al. 2015)

  • Consider implementing an expectation-maximization (EM)-based online CTC algorithm for sequence training of unidirectional RNNs, enabling them to learn sequences longer than the amount of unrolling and efficiently adapt to varying sequence lengths. (K. Hwang and Sung 2015)

  • Consider utilizing deep Long Short-Term Memory (LSTM) recurrent neural networks (RNNs) for speech recognition tasks, as they have demonstrated superiority over traditional feed-forward deep neural networks (DNNs), and can be further optimized through techniques like frame stacking, reduced frame rate, and context-dependent phone modeling. (Sak et al. 2015)

  • Employ a neural network architecture to effectively handle sparsity issues arising from integrating contextual information into classical statistical models, enabling them to develop dynamic-context generative models that consistently outperform both context-sensitive and non-context-sensitive Machine Translation and Information Retrieval baselines. (Sordoni et al. 2015)

  • Explore the potential benefits of using tree-structured LSTMs over traditional sequential LSTMs for improved semantic representations in various natural language processing tasks. (K. S. Tai, Socher, and Manning 2015)

  • Consider using a semantically controlled Long Short-Term Memory (LSTM) structure for your natural language generation (NLG) systems, as it allows for better optimization of sentence planning and surface realization, leading to more natural and varied language outputs. (T.-H. Wen et al. 2015)

  • Consider the “Goldilocks principle” when representing wider context in memory, finding an optimal size for memory representations between single words and entire sentences depending on the class of word being predicted. (Hill et al. 2015)

  • Utilize Gated Graph Sequence Neural Networks (GGS-NNs) for handling graph-structured data, as they provide a flexible and efficient approach for processing complex relationships within the data. (Yujia Li et al. 2015)

  • Carefully consider the benefits and limitations of recurrent neural networks (RNNs) compared to other models, such as Markov models, when working with sequential data, taking into account factors such as the ability to capture long-range time dependencies, computational feasibility, and the potential for overfitting. (Lipton, Berkowitz, and Elkan 2015)

  • Utilize the Eesen framework for end-to-end speech recognition, which employs deep recurrent neural networks (RNNs) and connectionist temporal classification (CTC) objective functions to simplify acoustic modeling, and uses weighted finite-state transducer (WFST) decoding to enable efficient incorporation of lexicons and language models. (Yajie Miao, Gowayyed, and Metze 2015)

  • Utilise both tree structure and sequence structure within Recurrent Neural Networks (RNNs) for superior performance in event extraction tasks., ‘The primary methodological insight presented in this study is the importance of incorporating both tree structure and sequence structure within Recurrent Neural Networks (RNNs) for optimal performance in event extraction tasks.’ (Mou et al. 2015)

  • Adopt a curriculum learning strategy to gradually transition from a fully guided training scheme using the true previous token to a less guided scheme primarily utilizing the generated token, thereby reducing the discrepancy between training and inference processes in sequence prediction tasks involving recurrent neural networks. (Vinyals, Kaiser, et al. 2014)

  • Consider utilizing knowledge transfer learning techniques to enhance the training of complex models like RNNs, leveraging the guidance of simpler models like DNNs, thereby achieving superior generalizability and performance. (Li Deng 2014)

  • Utilise a combination of deep convolutional neural networks (CNNs) and recurrent neural networks (RNNs) to develop a single joint model capable of accurately translating images into coherent, descriptive sentences. (Bahdanau, Cho, and Bengio 2014)

  • Focus on developing a generic tool for transforming an arbitrary st-graph into a feedforward mixture of RNNs, called structural-RNN (S-RNN), which can effectively capture complex spatio-temporal relationships while maintaining scalability. (L.-C. Chen et al. 2014)

  • Carefully evaluate the choice of recurrent units in recurrent neural networks, particularly considering more sophisticated options like LSTM and GRU, as they can significantly improve performance in tasks involving long-term dependencies. (J. Chung et al. 2014)

  • Consider implementing a recurrent neural network (RNN) model for attention-based task-driven visual processing, which enables the model to make decisions sequentially and incrementally build up a dynamic internal representation of the scene or environment, ultimately leading to improved efficiency and effectiveness in various applications. (V. Mnih et al. 2014)

  • Consider implementing a Clockwork Recurrent Neural Network (CW-RNN) architecture in your studies, as it demonstrates significant improvements in performance, reduced computational complexity, and faster evaluation times compared to traditional Simple Recurrent Neural Networks (SRNs) and Long Short-Term Memory (LSTM) networks. (Sak, Senior, and Beaufays 2014)

  • Consider using a multilayered Long Short-Term Memory (LSTM) to map input sequences to a fixed-dimensional vector, followed by another deep LSTM to decode the target sequence from the vector, as demonstrated by the authors successful application of this approach to English to French translation tasks.’ (Sutskever, Vinyals, and Le 2014)

  • Carefully apply dropout regularization to specific subsets of recurrent neural network connections to prevent overfitting and enhance performance across multiple tasks such as language modeling, speech recognition, image caption generation, and machine translation. (Zaremba, Sutskever, and Vinyals 2014)

  • Focus on developing a relevancy screening mechanism, inspired by cognitive processes, to efficiently consolidate relevant memory and achieve scalable use of sparse self-attention with recurrence in recurrent neural networks. (Alberini, Johnson, and Ye 2013)

  • Utilise deep learning algorithms to create synthetic benchmarks for predictive modeling purposes, rather than rely solely on traditional benchmark suites. (Graves 2013)

  • Explore the benefits of incorporating deep recurrent neural networks (DRNNs) in speech recognition tasks, as they effectively combine the advantages of deep networks with the ability of recurrent neural networks to utilize long-range context, leading to significant improvements in accuracy. (Graves, Mohamed, and Hinton 2013)

  • Utilise a variety of techniques including gradient clipping, leaky integration, advanced momentum techniques, more powerful output probability models, and encouragement of sparser gradients to overcome the challenges associated with learning long-term dependencies in recurrent neural networks. (Yoshua Bengio, Boulanger-Lewandowski, and Pascanu 2012)

  • Consider increasing the bias to the forget gate before attempting to use more sophisticated approaches in order to improve the performance of LSTMs. (Boulanger-Lewandowski, Bengio, and Vincent 2012)

  • Utilize a bidirectional dynamic multi-pooling long short-term memory tensor neural networks (BDLSTM-TNNs) for event extraction tasks, as it enables automatic induction of valuable clues without complex NLP preprocessing and simultaneous prediction of candidate arguments, thereby improving overall accuracy. (Zeiler 2012)

  • Consider using Tensor Train (TT) decomposition when attempting to compress recurrent neural networks while preserving your expressive power, as it outperforms other tensor decomposition methods like CANDECOMP/PARAFAC (CP) and Tucker decomposition in terms of performance on sequence modeling tasks. (Oseledets 2011)

  • Consider employing a Complex Evolutional Network (CEN) model to effectively capture the length-diversity and time-variability of evolutional patterns within Temporal Knowledge Graphs (TKGs) for accurate prediction of future facts. (Hosten et al. 2008)

  • Utilize a novel joint neural model for simultaneous entity recognition and relation extraction, which doesnt rely on any manually extracted features or external tools, thereby improving accuracy across diverse languages and contexts.’ (Y. Bengio, Simard, and Frasconi 1994)

  • Focus on understanding the dynamics of neural microcircuits from the perspective of a readout neuron, which can learn to extract salient information from the high-dimensional transient states of the circuit and transform transient circuit states into stable readouts, allowing for invariant readout despite the absence of revisiting the same state. (NA?)

  • Consider combining evolutionary algorithms with linear regression techniques to optimize the performance of recurrent neural networks, particularly in situations where gradient-based learning algorithms struggle due to rough error surfaces and numerous local minima. (NA?)

  • Consider using recurrent neural networks (RNNs) and echo state networks (ESNs) for malware classification tasks, as these models can effectively capture the “language” of malware and improve detection rates compared to traditional machine learning approaches. (NA?)

  • Utilise a flexible, gradient descent-based training of excitatory-inhibitory RNNs that can incorporate various forms of biological knowledge, especially regarding local and large-scale connectivity in the brain. (NA?)

  • Utilize an attention-based bilingual LSTM network for cross-lingual sentiment classification, which effectively models the compositional semantics and captures long-distance dependencies between words in bilingual texts. (NA?)

  • Consider utilizing deep convolutional neural networks (DCNNs) and long short-term memory (LSTM) recurrent neural networks together in a unified framework for human activity recognition (HAR) tasks, especially when working with multimodal wearable sensors. (NA?)

  • Consider using a fully data-driven, end-to-end trained neural sequence-to-sequence model with an encoder-decoder architecture consisting of two recurrent neural networks for performing retrosynthetic reaction prediction tasks, as it offers several advantages over traditional rule-based expert systems and hybrid deep learning approaches. (NA?)

  • Explore the use of deep neural networks and transfer learning for financial decision support, as your results show improved directional accuracy in predicting stock price movements in response to financial disclosures compared to traditional machine learning methods. (NA?)

  • Conduct large-scale analyses of different LSTM variants across diverse tasks, optimize hyperparameters separately for each task using random search, assess the importance of these hyperparameters using fANOVA, and draw conclusions about the efficiency and effectiveness of each LSTM variant based on these comprehensive evaluations. (NA?)

  • Utilise language models trained on correct source code to identify tokens that appear out of place, and subsequently consult those models to determine the most probable replacement tokens for the estimated error location. (NA?)

  • Consider using synthetic gradients to decouple neural network modules, enabling independent and asynchronous updates, thereby improving efficiency and flexibility in various applications. (NA?)

  • Utilise two novel neural architectures - one based on bidirectional LSTMs and conditional random fields, and the other that constructs and labels segments using a transition-based approach inspired by shift-reduce parsers - to achieve state-of-the-art performance in Named Entity Recognition (NER) across four languages without requiring any language-specific knowledge or resources such as gazetteers. (NA?)

  • Adopt deep recurrent neural networks (DRNNs) for human activity recognition tasks, specifically those involving variable-length input sequences, as these models are capable of capturing long-range dependencies and outperform conventional machine learning methods like SVM and KNN, as well as other deep learning techniques like DBNs and CNNs. (NA?)

  • Consider using a prompt-aware and attention-based LSTM-RNN model for scoring non-native spontaneous speech, as it outperforms traditional support vector regressors and does not require extensive feature engineering. (NA?)

  • Consider integrating user-behavioral data, such as tendencies toward racism or sexism, into your deep learning models for improved classification accuracy in detecting hate speech in social media posts. (NA?)

  • Utilize a joint neural model for simultaneous entity recognition and relation extraction, specifically modelling entity recognition through a Conditional Random Fields (CRF) layer and relation extraction as a multi-head selection problem, thereby avoiding reliance on external natural language processing (NLP) tools or manually extracted features. (NA?)

  • Employ a mixed neural network (MNN) approach combining rectifier neural network (RNN) and long short-term memory (LSTM) architectures to optimise classification performance in sleep stage classification tasks using single-channel EEG recordings. (NA?)

  • Focus on developing large-scale photonic Recurrent Neural Networks (RNNs) with numerous nonlinear nodes, utilizing reinforcement learning techniques to improve performance and energy efficiency. (NA?)

  • Carefully consider the choice of time resolutions when analyzing time series data, as different resolutions can reveal distinct patterns and improve overall prediction accuracy. (NA?)

  • Consider employing deep neural network architectures for detecting mental disorders like depression in social media platforms, particularly focusing on optimising word embeddings and comparing various deep learning architectures. (NA?)

  • Consider using a GRU-D model when dealing with missing values in time series data, as it incorporates trainable decay mechanisms that allow for improved utilization of missingness information compared to traditional imputation techniques. (NA?)

  • Carefully consider the selection of k-mer length, stride window, and embedding vector dimension when developing models for identifying transcription factor binding sites in DNA sequences, as these factors significantly impact model performance. (NA?)

  • Consider creating customized basecalling models using taxon-specific datasets and larger neural networks to achieve higher accuracy in basecalling tasks, while acknowledging the tradeoff between accuracy and processing speed. (NA?)

  • Consider utilizing Long Short-Term Memory networks (LSTMs) and Entity-Aware-LSTMs (EA-LSTMs) for regional rainfall-runoff modeling, as these techniques enable improved performance compared to traditional hydrological models and facilitate the learning of catchment similarities. (NA?)

  • Utilize a cascaded RNN model with GRUs for HSI classification, which effectively addresses the redundant and complementary information of HSIs through two RNN layers - one for reducing redundancy and the other for learning complementarity. (NA?)

  • Consider utilising deep neural networks when attempting to improve signal peptide predictions, as evidenced by the success of SignalP 5.0 in distinguishing between three types of prokaryotic signal peptides. (NA?)

  • Consider implementing physical reservoir computing systems using various physical phenomena as reservoirs, rather than relying solely on traditional recurrent neural networks, in order to achieve faster information processing and lower learning costs. (NA?)

  • Carefully consider the unique challenges posed by different types of entities when developing entity linking frameworks, and tailor your approach accordingly. (NA?)

  • Explore the potential of wave physics as an alternative to digital implementations for developing analog machine learning hardware platforms, due to its ability to passively process signals and information in your native domain, resulting in significant gains in speed and reductions in power consumption. (NA?)

  • Utilise multitask learning approaches when dealing with clinical time series data, as this enables simultaneous handling of various clinical prediction tasks, thereby improving overall model performance. (NA?)

  • Consider utilizing both shallow machine learning (XGBoost) and deep learning (LSTM) methods for building thermal load prediction, recognizing that each method may excel in different scenarios based on factors such as prediction horizon and input uncertainty. (NA?)

  • Use a combination of traditional statistical methods like the modified SEIR model and advanced techniques like machine learning algorithms to accurately predict the trajectory of infectious diseases like COVID-19. (NA?)

  • Ensure they fully understand the foundational principles of RNN and LSTM networks before attempting to implement them, as this will allow them to develop a deeper intuition for how these systems operate and avoid common pitfalls. (NA?)

  • Consider utilizing long short-term memory (LSTM) and convolutional neural networks (CNN) for time series forecasting, as they demonstrated superior performance in the study. (NA?)

Long Short-Term Memory (Lstm)

  • Consider using Extreme Value Loss (EVL) instead of conventional quadratic loss when dealing with time series prediction involving extreme events, and they may benefit from integrating a Memory Network to capture historical extreme events. (D. Ding et al. 2019)

  • Leverage emojis as an instrument to improve cross-lingual sentiment analysis by integrating language-specific representations and feeding them through downstream tasks to predict real, high-quality sentiment labels in the source language. (Zhenpeng Chen et al. 2019)

  • Consider using a bidirectional Long Short-Term Memory (LSTM) recurrent neural network for onset detection in music signals, as it offers superior performance and temporal precision compared to traditional methods. (Eyben 2016)

  • Use character-level language models as an interpretable testbed to understand the long-range dependencies learned by LSTMs, and compare your performance against (n)-gram models to identify areas for improvement. (Karpathy, Johnson, and Fei-Fei 2015)

  • Consider utilizing convolutional LSTM (ConvLSTM) networks for spatiotemporal sequence forecasting problems, as they demonstrate superior performance compared to fully connected LSTM (FC-LSTM) and existing operational algorithms in precipitation nowcasting. (X. Shi et al. 2015)

  • Consider implementing Dynamic Layer Normalization (DLN) in your neural acoustic models for speech recognition tasks, as it enables the model to dynamically adapt to variations in acoustics caused by differences in speakers, channels, and environments without requiring additional adaptation data or increasing model size. (Dieleman et al. 2015)

  • Consider utilizing Deep Belief Networks (DBNs) for feature extraction and classification tasks, as demonstrated through the DeeBNet V3.0 toolbox, which offers improved accuracy and flexibility across various domains such as image, speech, and text processing. (Keyvanrad and Homayounpour 2014)

  • Consider employing deep learning techniques, specifically deep belief networks and restricted Boltzmann machines, for improved feature learning and representation in neuroimaging studies. (Plis et al. 2014)

  • Consider using the Persistent Contrastive Divergence (PCD) algorithm for training Restricted Boltzmann Machines (RBMs) as it outperforms traditional Contrastive Divergence (CD) and Pseudo-Likelihood algorithms while maintaining similar speed and simplicity. (NA?)

  • Carefully choose the appropriate type of restricted Boltzmann machine (RBM) based on the specific characteristics of your dataset, and optimize various parameters such as learning rate, momentum, weight decay, and sparsity to ensure effective training and prevent overfitting. (NA?)

  • Consider utilizing semi-supervised anomaly detection methods, specifically the Discriminative Restricted Boltzmann Machine, to effectively analyze and classify network traffic while remaining adaptive to changing network environments. (NA?)

  • Prioritize topological sparsity in the ANN design phase, resulting in significantly reduced connections and improved memory and computational efficiency. (NA?)

  • Utilise machine learning techniques, specifically artificial neural networks, for quantum state tomography (QST) of highly-entangled states in both one and two dimensions. (NA?)

Deep Belief Networks (Dbn)

  • Utilize a higher-order Boltzmann machine that includes multiplicative interactions among groups of hidden units encoding distinct factors of variation, combined with correspondence-based training strategies, to effectively disentangle and model the joint interaction of various latent factors influencing sensory data. (Desjardins, Courville, and Bengio 2012)

  • Utilise the Sparse Encoding Symmetric Machine (SESM) algorithm for unsupervised learning tasks, as it effectively balances the trade-off between reconstruction error and information content of the representation, leading to improved accuracy and reduced computational complexity. (NA?)

  • Utilise the convolutional deep belief network (CDNB) model for scalable unsupervised learning of hierarchical representations, particularly in the field of computer vision. (NA?)

  • Utilize a combination of variational approximation and persistent Markov chains to efficiently estimate data-dependent and data-independent statistics, respectively, enabling the successful learning of complex Boltzmann machines. (NA?)

  • Utilize Deep Belief Networks (DBNs) for natural language understanding tasks, as they provide superior performance compared to traditional methods like Support Vector Machines (SVM), boosting, and Maximum Entropy (MaxEnt) when initialized with unsupervised pre-training and combined with original features. (NA?)

  • Consider utilizing deep learning techniques, specifically Deep Belief Networks, to enhance the performance of just-in-time defect prediction systems. (NA?)

  • Focus on developing improved training algorithms for restricted Boltzmann machines (RBMs) by analyzing the bias of contrastive divergence (CD) approximation, establishing bounds on the mixing rate of parallel tempering (PT), and exploring novel approaches like centered RBMs and estimation techniques from statistical physics to enhance the efficiency and effectiveness of RBM training. (NA?)

  • Consider using state representation learning (SRL) algorithms to create low-dimensional, interpretable, and action-influenced representations of complex environments, which can enhance the efficiency and effectiveness of downstream tasks like reinforcement learning and robotics control. (NA?)

  • Combine deep learning models with structured hierarchical Bayesian models to create compound HD (Hierarchical-Deep) models that can efficiently learn novel concepts from very few training examples by leveraging low-level generic features, high-level features that capture correlations among low-level features, and a category hierarchy for sharing priors over the high-level features that are typical of different kinds of concepts. (NA?)

Autoencoder

  • Investigate the effectiveness of unsupervised pre-training in deep learning models by conducting extensive simulations and testing multiple hypotheses, ultimately supporting the theory that unsupervised pre-training serves as a form of regularization that guides learning toward optimal solutions. (Taoli Cheng and Courville 2023)

  • Consider using a multi-scale masked autoencoder (Point-M2AE) for hierarchical self-supervised learning of 3D point clouds, as it effectively models spatial geometries and captures both fine-grained and high-level semantics of 3D shapes. (Renrui Zhang et al. 2022)

  • Consider both global and personal factors when analyzing heart rate time series data, as they interact and influence each other, leading to unique patterns within individuals. (Xian Wu et al. 2020)

  • Consider using deterministic autoencoders (RAEs) as a simpler, more scalable, and potentially superior alternative to traditional variational autoencoders (VAEs) for generative modeling tasks, particularly when dealing with high-dimensional data. (P. Ghosh et al. 2019)

  • Consider using a multiscale approach to generate high-resolution spectrograms in a coarse-to-fine order, which helps to overcome the bias of autoregressive models towards capturing local dependencies and improves overall audio fidelity. (Vasquez and Lewis 2019)

  • Consider using a Local-to-Global auto-encoder (L2G-AE) to improve your understanding of point clouds by simultaneously learning both local and global structures via local to global reconstruction, incorporating a hierarchical self-attention mechanism to emphasize significant points, scales, and regions at varying levels within the encoder. (Xinhai Liu et al. 2019)

  • Focus on developing an effective and efficient embedding algorithm that can quickly adapt to changing network structures and identify anomalies in real-time, while being scalable and requiring minimal computational resources. (W. Yu et al. 2018)

  • Carefully analyze the impact of noise on learning dynamics in denoising autoencoders, as it can lead to improved performance and faster training times. (Advani and Saxe 2017)

  • Consider utilizing a folding-based decoder within your deep auto-encoders for point cloud analysis, as it provides a highly effective and efficient means of transforming 2D grid data into 3D point cloud representations. (Achlioptas et al. 2017)

  • Consider using a WaveNet-style autoencoder model for audio synthesis, which conditions an autoregressive decoder on temporal codes learned from the raw audio waveform, and utilizes a large-scale, high-quality dataset like NSynth for training and evaluating the model. (J. Engel et al. 2017)

  • Utilize Point Auto-Encoder (PointAE) with skip-connection and attention block for 3D statistical shape and texture modelling directly on 3D points, allowing for improved correspondence refinement and simultaneous modelling of shape and texture variation. (Hyeongwoo Kim et al. 2017)

  • Consider incorporating neural networks into your collaborative filtering models to improve performance and address the cold start problem, particularly by utilizing stacked denoising autoencoders to capture non-linear relationships within the data. (Strub, Mary, and Gaudel 2016)

  • Consider utilizing the AutoRec framework when conducting collaborative filtering studies due to its superior performance compared to traditional methods such as biased matrix factorization, RBM-CF, and LLORMA, as demonstrated on the Movielens and Netflix datasets. (Sedhain et al. 2015)

  • Leverage the ability to generate images for the purpose of recognizing other images, utilizing a combination of hard-coded structures and learned content within a sophisticated autoencoder. (Yoshua Bengio et al. 2013)

  • Utilize denoising autoencoders to extract robust features from corrupted inputs, thereby improving the quality of your deep learning models. (NA?)

  • Consider incorporating a higher order contractive auto-encoder into your experimental designs, as it provides a more effective and computationally efficient method for unsupervised feature extraction compared to existing approaches. (NA?)

  • Utilize the conceptual linkage between denoising autoencoders and score matching to enhance your understanding of both approaches, thereby improving the efficiency and effectiveness of your statistical analyses. (NA?)

  • Consider combining stacked autoencoders (SAEs) with the extreme learning machine (ELM) to create an effective deep learning approach for accurately predicting building energy consumption. (NA?)

  • Consider using autoencoder networks to enable intuitive exploration of high-dimensional procedural modeling spaces within a lower dimensional space learned through autoencoder network training, allowing for faster and more efficient creation of high-quality content. (NA?)

  • Leverage deep learning algorithms to discover and represent eigenfunctions of the Koopman operator, allowing them to efficiently analyze and control nonlinear systems using linear theory. (NA?)

  • Consider utilizing multiple networks in your studies, rather than just focusing on individual networks, as it provides additional information and improves the overall quality of the findings. (NA?)

  • Consider utilizing autoregressive generative models for protein design and variant prediction, as they offer significant advantages over traditional alignment-based methods, especially for highly variable and diverse sequences like those found in antibodies. (NA?)

Variational Autoencoder (Vae)

  • Carefully examine potential linguistic biases in existing datasets before attempting to develop and evaluate models for ArtVQA, as demonstrated through the creation of the ArtQuest dataset. (A. Agrawal et al. 2022)

  • Consider utilising a Multi-Stage, Multi-Codebook (MSMC) approach to high performance neural Text-to-Speech (TTS) synthesis. This involves using a vector-quantized, variational autoencoder (VQ-VAE) based feature analyser to encode Mel spectrograms of speech training data by down-sampling progressively in multiple stages into MSMC Representations (MSMCRs) with different time resolutions, and quantizing (H. Guo et al. 2022)

  • Address the training-inference mismatch issue in unsupervised learning of controllable generative sequence models by employing a style transformation module to transfer target style information into an unrelated style input, enabling training using unpaired content and style samples. (“ESPnet2 Pretrained Model, Kamo-Naoyuki/Librispeech_asr_train_asr_conformer6_n_fft512_hop_length256_raw_en_bpe5000_scheduler_confwarmup_steps40000_optim_conflr0.0025_sp_valid.acc.ave, Fs=16k, Lang=en” 2021)

  • Consider using a variational auto-encoder based non-autoregressive text-to-speech (VAENAR-TTS) model for generating high-quality speech efficiently, as it eliminates the need for phoneme-level durations and provides a more flexible alignment between text and spectrogram. (Hui Lu et al. 2021)

  • Consider implementing a cyclical annealing schedule for Variational Autoencoders (VAEs) to address the KL vanishing issue, allowing for progressive learning of more meaningful latent codes and improved performance across a wide range of Natural Language Processing (NLP) tasks. (H. Fu et al. 2019)

  • Utilize automatic reparameterization techniques in probabilistic programming systems to optimize the efficiency and accuracy of inference algorithms, enabling robust inference across various models without requiring a priori knowledge of the optimal parameterization. (Gorinova, Moore, and Hoffman 2019)

  • Consider using the Inductive Topic Variational Graph Auto-Encoder (T-VGAE) model when dealing with text classification problems, as it effectively combines topic modelling and graph-based information propagation within a unified framework, providing improved interpretability and overall performance. (Lianzhe Huang et al. 2019)

  • Utilise a two-stage approach for generating diverse high-fidelity images: firstly, train a hierarchical VQ-VAE to encode images onto a discrete latent space, and subsequently, fit a powerful PixelCNN prior over the discrete latent space induced by all the data. (Razavi, Oord, and Vinyals 2019)

  • Adopt metric preservation as a powerful prior for learning latent representations of deformable 3D shapes, as it provides a rigorous way to control the amount of geometric distortion occurring in the construction of the latent space, leading to higher quality synthetic samples. (Chaudhuri, Ritchie, and Xu 2019)

  • Consider using a flow-based generative network called WaveGlow for speech synthesis tasks, as it provides fast, efficient, and high-quality audio synthesis without requiring autoregression, simplifying the training procedure and improving stability. (R. Yamamoto et al. 2018)

  • Consider leveraging the reparameterization trick to transform deep directed graphical models (DGMs) into a compact semi-auxiliary form, allowing for effective knowledge distillation without encountering intractability or error accumulation issues. (Achille et al. 2018)

  • Utilise the Temporal Difference Variational Auto-Encoder (TD-VAE) model for generating sequence models that meet specific criteria including building an abstract state representation, forming a belief state, and exhibiting temporal abstraction. (B. Amos et al. 2018)

  • Utilise a probabilistic fully-connected graph as the decoder output in a variational autoencoder to sidestep difficulties associated with linearisation of discrete graph structures. (Simonovsky and Komodakis 2018)

  • Utilize a straightforward variational Bayes scheme for Recurrent Neural Networks, which includes a simple adaptation of truncated backpropagation through time for better quality uncertainty estimates and superior regularization, while also demonstrating how a novel type of posterior approximation can enhance the performance of Bayesian RNNs. (Fortunato, Blundell, and Vinyals 2017)

  • Utilise a variational autoencoder to generate small graphs, particularly in the context of molecule generation, by outputting a probabilistic fully-connected graph of a predefined maximum size directly at once. (Goh et al. 2017)

  • Employ a syntax-directed variational autoencoder (SD-VAE) to improve the quality of your generative models for discrete structured data, such as computer programs and molecular structures, by ensuring both syntactic and semantic validity. (Benhenda 2017)

  • Utilise unsupervised boosting techniques to enhance the performance of generative models. (Grover and Ermon 2017)

  • Develop and evaluate adversarial attacks on deep generative models, such as Variational Autoencoders (VAEs) and VAE-Generative Adversarial Networks (VAE-GANs), to understand your vulnerability to malicious manipulations and improve your robustness. (Kos, Fischer, and Song 2017)

  • Adopt a Bayesian point of view in dealing with the issue of compression and computational efficiency in deep learning. They suggest using sparsity inducing priors to prune large parts of the network, thereby achieving state-of-the-art compression rates while maintaining competitiveness with other methods optimized for speed or energy efficiency. (Louizos, Ullrich, and Welling 2017)

  • Utilise a “Neural Statistician” model, which extends the variational autoencoder to learn a method for computing representations, or statistics, of datasets in an unsupervised manner. This allows for efficient learning from new datasets for both unsupervised and supervised tasks. (Edwards and Storkey 2016)

  • Consider implementing a “class-disentanglement” technique, which involves training a variational autoencoder to extract class-dependent information from an image, allowing for improved understanding of neural networks and enhanced detection and defense against adversarial attacks. (Alexander A. Alemi et al. 2016)

  • Consider using a cluster-wise hierarchical generative model for deep amortized clustering (CHiGac) to improve efficiency and accuracy in clustering datasets, as it enables simultaneous learning of cluster formation, data point grouping, and adaptive control of the number of clusters. (J. L. Ba, Kiros, and Hinton 2016)

  • Utilise Variational Autoencoders (VAEs) for unsupervised learning of complex distributions due to your ability to leverage standard function approximators (such as neural networks) and be trained efficiently with stochastic gradient descent. (Doersch 2016)

  • Consider using a combination of two convolutional network stacks - one that conditions on the current row and one that conditions on all rows above - to effectively eliminate the blind spot issue in the receptive field of the PixelCNN architecture, thereby enabling accurate and efficient image generation. (Oord, Kalchbrenner, et al. 2016)

  • Consider utilizing deep latent variable models for sequential data when dealing with complex, high-dimensional data sets, as these models offer a powerful and scalable solution for unsupervised learning. (Archer et al. 2015)

  • Consider using disentangled representation learning when working with unsupervised neural quantization to achieve better performance in non-exhaustive search applications. (Mirza and Osindero 2014)

  • Consider utilizing the multi-entity variational autoencoder (MVAE) model when attempting to learn object-based representations from data, as it demonstrates the ability to effectively disentangle objects and your properties in visual scenes. (Diederik P. Kingma and Welling 2013)

  • Consider using a regularization framework for variational autoencoders to ensure semantic validity in the generation of complex combinatorial structures like graphs. (Barabási and Albert 1999)

  • Utilise Variational Autoencoders (VAEs) for unsupervised learning tasks, particularly those involving complex systems or phase transitions, due to your capacity to effectively encode and recreate the original data, thus providing valuable insights into the systems behaviour.’ (NA?)

  • Consider using short-run MCMC, such as short-run Langevin dynamics, as an approximate flow-based inference engine for learning latent variable models, and correct the bias existing in the output distribution of the non-convergent short-run Langevin dynamics using optimal transport (OT) to improve the accuracy of the model parameter estimation. (NA?)

  • Consider employing variational autoencoders (VAEs) as a principled method for jointly learning deep latent-variable models and corresponding inference models using stochastic gradient descent, which offers numerous benefits across diverse applications such as generative modeling, semi-supervised learning, and representation learning. (NA?)

  • Focus on developing efficient and robust noisy decoder-based pseudo example generators for improved performance in semi-supervised learning and few-shot learning tasks. (NA?)

  • Focus on developing efficient and robust noisy decoder-based pseudo example generators for improved performance in semi-supervised learning (SSL) and few-shot learning (FSL) tasks. (NA?)

Generative Adversarial Networks (Gan)

  • Consider utilizing a graph-generative data augmentation framework called GraDA to enhance your commonsense reasoning datasets, as it effectively synthesizes factual data samples from knowledge graphs, leading to improved performance in various commonsense reasoning tasks. (Yu Chen, Wu, and Zaki 2024)

  • Consider using the MAGBIG benchmark to systematically assess and mitigate gender bias in multilingual text-to-image models, promoting inclusivity and fairness across diverse linguistic contexts. (Friedrich et al. 2024)

  • Develop a modular training algorithm for deep causal generative models that enables accurate sampling from identifiable interventional and counterfactual distributions, particularly when dealing with high-dimensional data such as images. (M. M. Rahman and Kocaoglu 2024)

  • Apply adversarial learning to in-context learning (ICL) to optimize the prompt for a given task, keeping model parameters fixed and updating the prompts in an adversarial manner, thus reducing computation and data requirements while enhancing model performance. (X. L. Do et al. 2023)

  • Consider utilizing various generative AI models for specific tasks, such as text-to-image, text-to-3D, image-to-text, text-to-video, text-to-audio, and text-to-code transformations, as these models offer unique advantages and potential applications across numerous industries. (Gozalo-Brizuela and Garrido-Merchan 2023)

  • Consider implementing prompt engineering techniques within a mobile-edge AIGX framework to optimize the quality of AI-generated content, enhance user satisfaction, and improve network performance. (Yinqiu Liu et al. 2023)

  • Explore the potential of natural phenomena, such as raindrops, as adversarial attackers to deep neural networks (DNNs), and develop techniques to generate adversarial raindrops using generative adversarial networks (GANs) to improve the robustness of DNNs to real-world raindrop attacks. (Jiyuan Liu et al. 2023)

  • Focus on developing a deep understanding of the specific requirements of large-scale text-to-image synthesis tasks, such as large capacity, stable training on diverse datasets, strong text alignment, and controllable variation vs. text alignment tradeoff, in order to optimize the performance of generative adversarial networks (GANs) in this domain. (Sauer et al. 2023)

  • Explore the potential of AI-generated content (AIGC) in various fields, considering its capabilities, limitations, and ethical implications, while focusing on the development of large-scale pre-trained models and integrating AIGC with metaverse applications. (Jiayang Wu et al. 2023)

  • Consider utilizing diffusion models for text-to-image tasks due to your ability to achieve high-quality image synthesis while maintaining strong alignment with the provided text. (Chenshuang Zhang, Zhang, Zhang, et al. 2023)

  • Consider utilizing the Gibbs zig-zag sampler, a novel combination of piecewise deterministic Markov processes (PDMPs) and Markov chain Monte Carlo (MCMC) techniques, to improve the efficiency and accuracy of statistical modeling in complex scenarios involving high-dimensional regression and random effects. (Sachs et al. 2023)

  • Utilize the HiFi++ framework when working on bandwidth extension and speech enhancement tasks, as it offers better or comparable performance to current state-of-the-art approaches while using significantly fewer computational resources. (Andreev et al. 2023)

  • Carefully phrase prompts to ensure accurate and reliable responses from GPT-3.5, taking into account sensitivity to wording and potential biases such as response order bias. (Aher, Arriaga, and Kalai 2022)

  • Utilise DATID-3D, a domain adaptation method specifically designed for 3D generative models, to effectively adapt these models across various domains while maintaining diversity and improving text-image correspondence. (Alanov, Titov, and Vetrov 2022)

  • Consider using prompt tuning for transfer learning of generative transformers, as it enables efficient adaptation to new domains and significantly improves image generation quality compared to traditional approaches. (Bahng et al. 2022)

  • Consider using Generative Adversarial CLIPs (GALIP) for text-to-image synthesis because it offers improved accuracy, reduced training time and data requirements, and enhanced controllability compared to existing methods. (Balaji et al. 2022)

  • Carefully examine the extent of content replication in diffusion models, especially those trained on large datasets, to ensure proper attribution and avoid potential legal issues. (Bardes, Ponce, and LeCun 2022)

  • Consider employing a stack of time-aware location-variable convolutions of diverse receptive field patterns to efficiently model long-term time dependencies with adaptive conditions, along with a noise schedule predictor to reduce the sampling steps without compromising the generation quality, particularly in the context of speech synthesis. (R. Huang et al. 2022)

  • Consider integrating source-filter modeling into your HiFi-GAN framework to achieve both fast synthesis and high F0 controllability in your neural vocoder designs. (Yoneyama, Wu, and Toda 2022)

  • Consider using a combination of adversarial training strategies and multi-singer conditional discriminators to optimize your singing voice synthesis systems, resulting in more natural and realistic singing voices. (Zewang Zhang et al. 2022)

  • Consider using conditional generative adversarial networks (cGANs) to create synthetic data for handwritten text recognition tasks, as this approach allows for greater control and flexibility in generating images from different given types compared to traditional methods. (L. Kang et al. 2022)

  • Consider using an unsupervised conditional GAN-based approach for generating Neural Radiance Fields (NeRF) from a single image, without requiring 3D, multi-view, or pose supervision. (Obukhov et al. 2021)

  • Focus on separating emotional features from emotion-independent features during emotional voice conversion tasks to enhance voice quality and achieve successful data augmentation. (Xiangheng He et al. 2021)

  • Utilize a multi-resolution spectrogram discriminator when working with neural vocoders to enhance the spectral resolution of waveforms and mitigate the over-smoothing issue. (W. Jang et al. 2021)

  • Consider adopting the StarGAN v2 framework for unsupervised non-parallel many-to-many voice conversion tasks, as it significantly outperforms previous models in producing natural-sounding voices and can generalize to a wide range of voice conversion scenarios. (Y. A. Li, Zare, and Mesgarani 2021)

  • Consider using proxy distributions, specifically those derived from diffusion-based generative models, to enhance the adversarial robustness of deep neural networks, as they have shown significant improvements in performance across various datasets and threat models. (Sehwag et al. 2021)

  • Consider utilizing Generative Adversarial Network (GAN) inversion for unsupervised 3D shape completion tasks, as it allows for greater generalization capabilities and avoids the need for paired training data. (Junzhe Zhang et al. 2021)

  • Utilise the iBOT framework for masked image modelling (MIM) because it allows for self-distillation on masked patch tokens and class tokens, enabling the online tokeniser to be jointly learnable with the MIM objective, thereby eliminating the need for a multi-stage training pipeline where the tokeniser must be pre-trained beforehand. (Jinghao Zhou et al. 2021)

  • Utilise the Physics Informed Discriminator (PID)-GAN framework over the existing Physics-Informed Generator (PIG)-GAN framework for uncertainty quantification tasks in deep learning. This is because the PID-GAN framework effectively addresses the issue of imbalanced generator gradients and fully leverages the potential of the adversarial optimization process inherent in GAN-based frameworks for minimizing complex physics-based loss functions. Furthermore, unlike the PIG-G (Daw, Maruf, and Karpatne 2021)

  • Utilize two distinct regularization strategies to prevent mode collapse in deep SVDD: one based on random noise injection through the standard cross-entropy loss, and another that penalizes mini-batch variance when it drops below a specific threshold. Additionally, they suggest implementing an adaptive weighting system to manage the balance between the SVDD loss and the corresponding regularizer. (Chong et al. 2020)

  • Consider implementing a Double Oracle Framework for Generative Adversarial Networks (DO-GAN) to efficiently compute mixed Nash equilibria in large-scale games, improving upon traditional methods by incorporating a linear program to find the exact mixed Nash equilibrium in polynomial time. (Farnia and Ozdaglar 2020)

  • Consider using Markov chain Monte Carlo (MCMC) methods for analyzing complex Bayesian models, as they create sequences of dependent variables that converge to the distribution of interest, making them robust and universally applicable, despite your limitations in terms of reaching stationarity and dealing with correlation among the variables. (Robert and Changye 2020)

  • Consider using an adversarial data augmentation framework comprising a generator, a discriminator, and an auxiliary discriminator to improve the performance of risk assessment models in cases where there is a significant class imbalance issue. (Yang Liu et al. 2020)

  • Consider using Style-Adaptive Layer Normalization (SALN) in conjunction with meta-learning techniques to enhance the performance of text-to-speech systems, particularly in cases involving few-shot generation and classification. (Karras, Laine, and Aila 2019)

  • Focus on developing models that combine the strengths of Generative Adversarial Networks (GANs) and Transformer architectures, specifically by creating a bipartite structure that enables long-range interactions across the image while maintaining computation of linear efficiency, ultimately improving the quality and diversity of generated images. (Bello et al. 2019)

  • Consider using a latent overcomplete GAN (LOGAN) for unpaired shape-to-shape translation, as it enables implicit feature disentanglement and adaptability to various types of transformations, such as content and style transfers, without requiring architectural modifications or parameter adjustments. (K. Yin et al. 2019)

  • Utilise a combination of Denoising Autoencoder networks (DAE) and Graph Neural Networks (GNN) to effectively generate classification weights for few-shot learning tasks. (Gidaris and Komodakis 2019)

  • Consider using a combination of convolutional neural networks (CNNs) and long short-term memory (LSTM) networks to improve the accuracy of your predictions in the field of handwritten text analysis. (B. Ji and Chen 2019)

  • Consider utilising a combination of adversarial, uniform, and reconstruction losses in order to optimise the performance of your generative adversarial network (GAN) models, specifically in the field of point cloud upsampling. (Ruihui Li et al. 2019)

  • Consider implementing a two-level domain confusion scheme within your adversarial learning objective, whereby the category-level confusion loss drives the learning of intermediate network features to be invariant at the corresponding categories of the two domains, thereby enhancing overall domain-invariant feature learning. (Yabin Zhang et al. 2019)

  • Use a Multiple-Objective Generative Adversarial Active Learning (MO-GAAL) approach instead of a Single-Objective Generative Adversarial Active Learning (SO-GAAL) approach for outlier detection tasks, because MO-GAAL prevents the generator from falling into the mode collapsing problem and generates a mixture of multiple reference distributions for the entire dataset. (Yezheng Liu et al. 2019)

  • Focus on developing a purely data-driven semi-supervised anomaly detection method based on the analysis of the hidden activations of neural networks, which they refer to as A^3. (“Computer Vision – ACCV 2018” 2019)

  • Consider using a generative adversarial network (GAN) architecture consisting of a generator and a discriminator, with the generator incorporating two layers of bidirectional long short-term memory (BiLSTM) networks and a dropout layer, and the discriminator being built upon a convolutional neural network (CNN), to effectively learn from existing ECG data and generate new ECGs that closely resemble the distribution of the original data. (F. Zhu et al. 2019)

  • Use adversarial training (AdvT) as a regularization method for network embedding models to enhance your robustness and generalization abilities, particularly by generating adversarial perturbations in the embedding space rather than the discrete graph domain. (Q. Dai et al. 2019)

  • Utilise the “instance-aware GAN” (InstaGAN) methodology for improved accuracy in image-to-image translation tasks, particularly those involving multiple target instances and significant shape changes. (Almahairi et al. 2018)

  • Utilise adversarial network compression techniques to transfer knowledge from a larger, more complex deep network to a smaller, less complex one, thereby improving the efficiency and effectiveness of the smaller network without compromising its performance. (Belagiannis, Farshad, and Galasso 2018)

  • Utilise the Cross-Domain Adversarial Auto-Encoder (CDAAE) model for effective domain adaptation in scenarios involving unlabelled data. (H. Hou, Huo, and Gao 2018)

  • Utilise a balancing generative adversarial network (BAGAN) to restore balance in imbalanced datasets, which involves incorporating all available images of majority and minority classes during adversarial training, allowing the generative model to learn useful features from majority classes and use these to generate images for minority classes. (Mariani et al. 2018)

  • Carefully examine the stability of your GAN training algorithms, particularly when dealing with data distributions that are concentrated on lower dimensional manifolds, as instability can arise due to discriminator gradients being orthogonal to the data distribution. (Mescheder, Geiger, and Nowozin 2018)

  • Consider using generative adversarial networks (GANs) to generate adversarial examples for deep neural networks (DNNs), as this approach can lead to more perceptually realistic examples and potentially accelerate adversarial training as defenses. (C. Xiao et al. 2018)

  • Consider using a combination of Autoencoders (AEs) and Generative Adversarial Networks (GANs) in the latent space for generating high-quality point clouds with improved fidelity and coverage of the original data. (Achlioptas et al. 2017)

  • Focus on understanding and leveraging the relationship between adversarial examples and the training distribution, specifically by identifying and mitigating the impact of low probability regions in the training distribution on the performance of machine learning models. (Yang Song et al. 2017)

  • Consider using Location-Aware Generative Adversarial Networks (LAGANs) for generating realistic radiation patterns from simulated high energy particle collisions, as they effectively capture the desired low-dimensional physical properties and offer a foundation for faster simulation in High Energy Particle Physics. (Paganini 2017)

  • Consider utilising knowledge distillation techniques to effectively compress Generative Adversarial Networks (GANs) for deployment in low SWAP (Size, Weight, and Power) hardware environments, such as mobile devices, while maintaining the quality of the generated output. (Yim et al. 2017)

  • Consider implementing network pruning during GANs training to explore different sub-network structures, thereby reducing the risk of prematurely pruning important connections and improving overall training efficiency. (X. Mao et al. 2017)

  • Consider using latent-space GANs (l-GANs) for generating point clouds because they are easier to train than raw GANs, achieve superior reconstruction, and offer better coverage of the data distribution. (Achlioptas et al. 2017)

  • Focus on developing a deep-learning approach to photographic style transfer that effectively combines structure preservation and semantic accuracy, resulting in photorealistic style transfers that maintain the integrity of the original image content. (F. Luan et al. 2017)

  • Use GraphGAN, a novel graph representation learning framework that combines generative and discriminative models through a game-theoretical minimax game, resulting in improved performance across multiple applications such as link prediction, node classification, and recommendation. (Hongwei Wang et al. 2017)

  • Consider utilizing generative adversarial networks (GANs) for various applications due to your ability to effectively handle complex, high-dimensional probability distributions, generate realistic samples, and adapt to diverse scenarios. (I. Goodfellow 2017)

  • Utilize Generative Adversarial Networks (GANs) for anomaly detection in high-dimensional data, as it provides a robust and effective solution for identifying unusual patterns within complex datasets. (Arjovsky and Bottou 2017)

  • Utilise a three-player game approach, namely KDGAN, instead of traditional two-player games like GAN, to effectively train a lightweight classifier for multi-label learning tasks. This approach allows the classifier to learn the true data distribution at the equilibrium, thereby increasing its accuracy and efficiency. (Arjovsky, Chintala, and Bottou 2017)

  • Consider incorporating the concept of Complementary Attention Feature (CAFE) in your Generative Adversarial Network (GAN) models to effectively edit only the parts of a face pertinent to the target attributes, thereby avoiding unintended alterations in facial regions. (Arjovsky, Chintala, and Bottou 2017)

  • Utilise a novel equilibrium enforcing method paired with a loss derived from the Wasserstein distance for training auto-encoder based Generative Adversarial Networks. This approach ensures a balance between the generator and discriminator during training, providing a new approximate convergence measure, faster and more stable training, and superior visual quality. (Berthelot, Schumm, and Metz 2017)

  • Use a novel end-to-end method called “Face Conditional Generative Adversarial Network” (FCGAN) to learn the mapping between low-resolution single face images and high-resolution ones, resulting in improved peak signal-to-noise ratio (PSNR) and overall visual quality. (Bin et al. 2017)

  • Leverage the power of deep generative adversarial training, specifically conditional generative adversarial networks, to address the cross-modal audio-visual generation problem, focusing on both instrument-oriented and pose-oriented generation scenarios. (L. Chen et al. 2017)

  • Utilise the Text Conditioned Auxiliary Classifier Generative Adversarial Network (TAC-GAN) when aiming to create high-quality, diverse, and discriminable images from text descriptions. (Dash et al. 2017)

  • Extend OpenMax by incorporating generative adversarial networks (GANs) for novel category image synthesis in order to explicitly model and provide decision scores for unknown classes in multi-class open set classification. (Z. Ge et al. 2017)

  • Utilize a Generative Adversarial Network (GAN) instead of traditional rule-based methods for password guessing tasks, as demonstrated by the superior performance of PassGAN in generating high-quality password guesses without requiring any a-priori knowledge about passwords or common password structures. (Hitaj et al. 2017)

  • Utilise a combination of cycle-consistency and semantic losses to maintain local structural information and semantic consistency when conducting unsupervised domain adaptation. (J. Hoffman et al. 2017)

  • Explore the higher-level parameter space for Neural Style Transfer and find a set of working shortcuts to map them to a reduced but meaningful set of creative controls. (B. Joshi, Stewart, and Shapiro 2017)

  • Utilise a novel approach called “DiscoGAN” to effectively discover cross-domain relations without requiring expensive pairing or extensive labelling. (T. Kim et al. 2017)

  • Utilise a novel approach called “DiscoGAN” to effectively discover cross-domain relations in unpaired data, thereby enabling successful transfer of style from one domain to another while preserving key attributes. (T. Kim et al. 2017)

  • Consider using a novel framework of cycle-consistent generative adversarial networks for unsupervised learning in style transfer problems involving asymmetric functions, such as makeup application and removal. (J. Liao et al. 2017)

  • Utilise a Generative Adversarial Network (GAN)-based model to transform source-domain images into appearing as if they were sampled from the target domain. This approach provides several benefits including decoupling from the task-specific architecture, generalisation across label spaces, improved training stability, potential for data augmentation, and interpretability. (Bousmalis et al. 2016)

  • Consider utilizing Plug and Play Generative Networks (PPGNs) for improved image generation, as they offer a flexible and adaptable framework that enables the creation of high-quality, diverse images through the combination of a generator network and a replaceable condition network. (Creswell, Arulkumaran, and Bharath 2016)

  • Utilise a Poisson process model to unify the perturbation and accept-reject views of Monte Carlo simulation, thereby enabling analysis of various methods such as A* sampling and OS*. (Maddison 2016)

  • Consider using Least Squares Generative Adversarial Networks (LSGANs) instead of regular GANs due to its ability to generate higher quality images and provide greater stability during the learning process. (X. Mao et al. 2016)

  • Utilise the Auxiliary Classifier GAN (AC-GAN) model for image synthesis, which incorporates both class-conditionality and an auxiliary decoder for reconstructing class labels, leading to improved sample quality and stability in training. (Mohamed and Lakshminarayanan 2016)

  • Utilise the auxiliary classifier GAN (AC-GAN) model for image synthesis, which incorporates both class-conditionality and an auxiliary decoder for reconstructing class labels, leading to improved sample quality and stability in training. (Odena, Olah, and Shlens 2016)

  • Consider utilizing a combination of deep convolutional generative adversarial networks (GANs) and recurrent neural network architectures to effectively translate visual concepts from characters to pixels, enabling the automatic synthesis of realistic images from text. (S. Reed et al. 2016)

  • Consider utilizing a topological GAN loss to ensure that your synthetic images accurately represent the topological features present in real images, thereby improving the overall accuracy and effectiveness of your downstream analyses. (Abbasi-Sureshjani et al. 2016)

  • Focus on developing regularizers for Generative Adversarial Networks (GANs) to address issues of training instability and missing modes, thereby improving the performance and reliability of these models. (J. Donahue, Krähenbühl, and Darrell 2016)

  • Consider using an energy-based Generative Adversarial Network (EBGAN) model, which treats the discriminator as an energy function that associates lower energies with regions close to the data manifold and higher energies elsewhere. This approach allows for increased flexibility in terms of architecture and loss functions, and can lead to more stable training behavior compared to traditional GANs. (Junbo Zhao, Mathieu, and LeCun 2016)

  • Consider using MelGAN, a non-autoregressive feed-forward convolutional architecture, for efficient and effective audio waveform generation in a GAN setup, as it yields high-quality text-to-speech synthesis models without requiring additional distillation or perceptual loss functions. (MORISE, YOKOMORI, and OZAWA 2016)

  • Consider implementing a self-regulating learning approach using a generative adversarial network to identify and remove spurious features in event detection tasks, thereby improving overall accuracy and adaptability. (X. Feng et al. 2016)

  • Utilise a combination of a multi-class GAN loss, an f-preservation component, and a regularisation component that encourages G to map samples from T to themselves, in order to effectively transfer a sample from one domain to an analogous sample in another domain. (Brock et al. 2016)

  • Consider integrating semantic annotation into your generative architectures to improve the predictability and quality of outputs, especially in areas like image synthesis and style transfer. (Champandard 2016)

  • Utilise optimal transport for feature alignment between conditional inputs and style exemplars in image translation, as it mitigates the constraint of many-to-one feature matching significantly while building up accurate semantic correspondences between conditional inputs and exemplars. (Chizat et al. 2016)

  • Leverage the power of context-conditional generative adversarial networks (CC-GANs) for semi-supervised learning, particularly in scenarios where there is a scarcity of labeled data. (Denton, Gross, and Fergus 2016)

  • Consider integrating efficient inference with the GAN framework through the development of an adversarially learned inference (ALI) model, which involves casting the learning of both an inference machine (or encoder) and a deep directed generative model (or decoder) within a GAN-like adversarial framework. (Dumoulin et al. 2016)

  • Consider using a generative adversarial network (GAN) based approach for imitation learning, as it enables them to directly extract a policy from data without going through the intermediate steps of inverse reinforcement learning, leading to improved performance in complex, high-dimensional environments. (Ho and Ermon 2016)

  • Utilise conditional adversarial networks (cGANs) as a general-purpose solution for image-to-image translation problems. This approach enables the network to learn the mapping from input image to output image, as well as the loss function required to train this mapping. By doing so, the same generic approach can be applied to various problems that typically demand distinct loss formulations. (Isola et al. 2016)

  • Consider using Markovian Generative Adversarial Networks (MGANs) for efficient texture synthesis, as it enables rapid generation of high-quality textures while reducing computational costs compared to previous methods. (Chuan Li and Wand 2016)

  • Consider implementing various techniques to enhance the stability and efficiency of Generative Adversarial Networks (GANs) training, including feature matching, minibatch discrimination, historical averaging, one-sided label smoothing, and virtual batch normalization. (Salimans et al. 2016)

  • Consider utilizing GPU-based parallel computing to speed up computations involving nearest-neighbor loss functions, as demonstrated through efficient implementation of Eq. 8 in the main paper. (L. Zheng, Yang, and Hauptmann 2016)

  • Focus on understanding how generative adversarial networks (GANs) work, your advantages and limitations, and explore ways to combine them with other methods to enhance performance and address challenges such as mode collapse. (Isola et al. 2016)

  • Consider using a recurrent text-to-image GAN when dealing with sequential data, as it enables accurate color rendering and improved consistency across image sequences compared to traditional text-to-image GANs. (Shaoqing Ren et al. 2015)

  • Consider using Pareto smoothed importance sampling (PSIS) to stabilize your importance sampling estimates, especially when dealing with high dimensional data, as it offers better performance than traditional methods like truncated importance sampling (TIS) and allows for accurate estimation of the Monte Carlo standard error (MCSE) and effective sample size (ESS). (Vehtari et al. 2015)

  • Consider implementing various techniques to improve the stability and convergence of Generative Adversarial Networks (GANs), including feature matching, minibatch discrimination, historical averaging, one-sided label smoothing, and virtual batch normalization, in order to enhance your ability to generate high-quality synthetic data. (Denton et al. 2015)

  • Consider using a dedicated GAN-based approach with unpaired image sets for training, along with two simple yet effective loss functions - a semantic content loss and an edge-promoting adversarial loss - to effectively learn the mapping from real-world photos to cartoon images, producing high-quality stylized cartoons that significantly outperform state-of-the-art methods. (Gatys, Ecker, and Bethge 2015)

  • Use the Maximum Mean Discrepancy (MMD) technique from statistical hypothesis testing to simplify the training of generative adversarial networks (GANs) by transforming the difficult minimax optimization problem into a straightforward loss function that can be optimized using backpropagation. (Hao Fang et al. 2014)

  • Consider utilizing a combination of deep convolutional and recurrent neural networks to create a generative adversarial network (GAN) for effective translation of textual descriptions into realistic images. (Mirza and Osindero 2014)

  • Consider using a multi-level statistics transfer model for self-driven person image generation, allowing for flexible manipulation of person appearance and pose properties without requiring paired source-target images during training. (Diederik P. Kingma and Ba 2014)

  • Utilise the adversarial nets framework for modelling complex distributions, as it offers superior performance compared to traditional methods due to its ability to generate diverse samples without requiring explicit representations of the underlying distribution, relying solely on backpropagation for gradient calculation, and eliminating the need for Markov chains or inference during learning. (I. J. Goodfellow et al. 2014)

  • Use conditional adversarial domain adaptation (CDAN) to improve the performance of deep networks in domain adaptation tasks, particularly when dealing with complex multimodal distributions. (Mirza and Osindero 2014)

  • Consider using latent subspace optimization when working with few-shot image generation problems, as it has been demonstrated to achieve superior performance in terms of diversity and generation quality compared to existing approaches. (Mirza and Osindero 2014)

  • Consider employing a Collaborative and Adversarial Network (CAN) for unsupervised domain adaptation, which involves training neural networks through domain-collaborative and domain-adversarial learning to achieve both domain-invariant and discriminant representations for improved image classification. (Tzeng et al. 2014)

  • Focus on developing an iterative algorithm that generates samples from a given density on a manifold based solely on the ability to evaluate the function defining the manifold, rather than relying on derivative information or random walks. (Oh et al. 2013)

  • Utilize a novel framework called Generative Adversarial Networks’, which uses a competitive relationship between two models - a generative model and a discriminative model - to estimate complex data distributions.’ (I. J. Goodfellow, Warde-Farley, Lamblin, et al. 2013)

  • Utilise full-batch Hamiltonian Monte Carlo (HMC) to accurately sample from the posterior distribution of Bayesian neural networks, despite its computational intensity, in order to gain deeper insight into the properties of these networks. (S. Ahn, Korattikara, and Welling 2012)

  • Utilize a Bayesian nonparametric approach to hidden Markov modeling, specifically through the implementation of a hierarchical Dirichlet process (HDP), to address the issue of unknown state numbers in the context of speaker diarization tasks. (E. B. Fox et al. 2011)

  • Utilise the proposed additive Gaussian processes model when dealing with regression tasks, as it offers improved interpretability and predictive power due to its ability to decompose functions into a sum of low-dimensional functions, each dependent on a subset of input variables. (Duvenaud, Nickisch, and Rasmussen 2011)

  • Consider employing Bayesian optimization techniques when dealing with expensive cost functions, as it enables them to balance exploration and exploitation effectively, thereby reducing the number of function evaluations needed. (Brochu, Cora, and Freitas 2010)

  • Consider adopting plug-and-play inference techniques for analyzing complex time series data, particularly when dealing with implicit models that do not provide explicit expressions for transition probabilities or sample paths. (Bretó et al. 2009)

  • Utilise a simulation-based methodology to verify the accuracy of software used to fit Bayesian models. (Cook, Gelman, and Rubin 2006)

  • Carefully consider the impact of crossmodal grounding shift when developing algorithms for low-resource adaptation of co-speech gesture generation models, as it can lead to significant improvements in performance. (Cassell, Vilhjálmsson, and Bickmore 2001)

  • Carefully consider the choice of Markov chain Monte Carlo (MCMC) algorithm, pay attention to convergence diagnostics, and utilize techniques such as reparameterization, blocking, collapsing, and cycling through different MCMC algorithms to improve mixing and ensure accurate estimation of posteriors. (Kass et al. 1998)

  • Utilise a Bayesian modelling approach when studying human concept learning, particularly when dealing with limited positive examples, as it offers superior explanatory power compared to alternative methods. (Feldman 1997)

  • Utilise a Bayesian adaptive psychometric method called QUEST, which uses prior knowledge and data to efficiently estimate the threshold of a psychometric function by placing trials at the current most probable estimate of threshold. (A. B. Watson and Pelli 1983)

  • Focus on developing algorithms capable of efficiently learning distributions generated by Probabilistic Stuffix Automata (PSAs), which can effectively approximate complex sequences with varying memory lengths, while maintaining computational efficiency. (NA?)

  • Utilise a probabilistic kernel approach to preference learning based on Gaussian processes, which offers a new likelihood function to capture preference relations within a Bayesian framework. (NA?)

  • Extend the differential evolution Markov chain (DE-MC) algorithm with a snooker updater, allowing them to effectively utilize fewer parallel chains while maintaining accuracy and efficiency in complex models. (NA?)

  • Employ a bottom-up ethnographic approach, combining an online questionnaire and an analysis of a large collection of user-generated prompts, to comprehensively understand the motivations, challenges, and usage patterns of text-to-image (TTI) practitioners. (NA?)

  • Utilise a Bayesian approach to model the physical characteristics of a star like α Cen A, employing a Markov chain Monte Carlo (MCMC) algorithm to estimate the posterior probability densities of the stellar parameters. This method becomes increasingly efficient relative to traditional grid-based strategies as the number of parameters increases, allowing for more accurate and robust estimates of the stellar parameters. (NA?)

  • Utilise tensor decompositions for learning latent variable models, specifically focusing on the extraction of a certain (orthogonal) decomposition of a symmetric tensor derived from the moments, which can be seen as a natural generalisation of the singular value decomposition for matrices. (NA?)

  • Consider utilizing Generative Adversarial Networks (GANs) under the constraint of differential privacy when attempting to create synthetic data sets for sharing purposes, as this approach offers a formal privacy guarantee and enables the creation of new plausible individuals without revealing sensitive information about any single study participant. (NA?)

  • Focus on finding new pseudo-words in the textual embedding space of pre-trained text-to-image models to effectively generate personalized text-to-image outputs without compromising the rich textual understanding and generalization capabilities of the model. (NA?)

  • Consider employing Generative Adversarial Networks (GANs) alongside metamorphic testing techniques to generate diverse and realistic driving scenes for testing the consistency and robustness of deep neural network-based autonomous driving systems. (NA?)

  • Utilize the table-GAN method when dealing with data privacy concerns, as it offers a balance between privacy protection and model compatibility through the use of generative adversarial networks (GANs) to synthesize fake tables that are statistically similar to the original table, thereby avoiding information leakage. (NA?)

  • Use a combination of deep learning techniques, specifically convolutional neural networks (CNNs) and conditional generative adversarial networks (cGANs), to accurately predict near-optimal topological designs without requiring any iterative schemes. (NA?)

  • Consider using conditional generative neural networks for global optimization tasks, as they can efficiently output ensembles of highly efficient topology-optimized metasurfaces operating across a range of parameters. (NA?)

  • Aim to develop efficient and stable deep learning algorithms for anomaly detection in multivariate time series, balancing accuracy with energy consumption and scalability concerns. (NA?)

  • Consider utilizing deep generative models for precipitation nowcasting, as they offer improved forecast quality, consistency, and value through producing realistic and spatiotemporally consistent predictions over large regions and lead times. (NA?)

  • Carefully evaluate the interplay between continuous and discrete state spaces when exploring the design space of E(3)-equivariant diffusion models for de novo 3D molecule generation, considering factors such as time-dependent loss weighting, inclusion of chemically motivated additional features, and transferability to different data distributions. (NA?)

  • Consider employing explainable artificial intelligence (XAI) techniques to enhance the interpretability and effectiveness of your text-to-image generative models, particularly in the context of emotional expression. (NA?)

Transformer Architecture

  • Carefully balance watermark robustness and text quality when developing watermarking techniques for large language models, considering factors such as sentence entropy and the impact of watermarking on the performance of pretrained models. (Baldassini et al. 2024)

  • Consider implementing Bi-directional Tuning for Lossless Acceleration (BiTA) in large language models (LLMs) to significantly improve your inference efficiency without sacrificing model performance. (F. Lin et al. 2024)

  • Consider using a prompt-based approach like CodePrompt when working on source code-related classification tasks, as it allows for the efficient retrieval of abundant knowledge from a language model, reduces computational costs, and increases overall task accuracy. (Yong Ma et al. 2024)

  • Adopt a unified mathematical framework to analyze different types of neural language models, enabling a deeper understanding of your inner workings and motivations behind your architectures. (M. Saleh and Paquelet 2024)

  • Focus on developing methods that leverage the discrepancy between the output distribution of large language models (LLMs) and the input-output mappings of a given task to improve the efficiency and effectiveness of in-context learning (ICL) demonstration selection. (S. Xu and Zhang 2024)

  • Consider employing speculative retrieval with batched verification to accelerate iterative RaLM serving while preserving model outputs. (Zhihao Zhang et al. 2024)

  • Consider using a dual attention framework to align the learning and selection processes for parameter-efficient tuning (PET) blocks in large language models, allowing for simultaneous handling of catastrophic forgetting and knowledge transfer challenges. (Weixiang Zhao et al. 2024)

  • Develop domain-specific models like medBERT.de for optimal performance in handling specialized text data, such as medical documents, and that data deduplication and efficient tokenization play only minor roles in enhancing model performance. (Bressem et al. 2024)

  • Consider combining large language models (LLMs) and knowledge graphs (KGs) in order to create a more robust and accurate system for natural language processing and artificial intelligence tasks. (S. Pan et al. 2024)

  • Consider leveraging the contradiction knowledge-enhanced prompts to tune the PLMs for improved sarcasm recognition. (Xueqi Cheng et al. 2023)

  • Carefully consider the impact of source-reference divergence in data collection and the effects of imperfect representation learning during training and inference when developing NLG models to reduce hallucination. (Z. Ji et al. 2023)

  • Carefully examine the impact of different preprocessing techniques on the training data, as some may unexpectedly degrade performance, such as filtering files from repositories with 5+ GitHub stars. (Allal et al. 2023)

  • Consider the impact of bounded entries on the efficiency of attention computation in large language models, as demonstrated by the authors investigation into the existence of faster algorithms through implicit usage of the matrix A.’ (Alman and Song 2023)

  • Consider using the tuned lens’ method for analyzing transformer models, as it provides more accurate, reliable, and unbiased predictions compared to the previously used ‘logit lens’. (Belrose et al. 2023)

  • Employ a fine-tuned Large Language Model (LLM) to distill question-answer pairs from raw sources, followed by LLM fine-tuning, in order to effectively retrieve knowledge, generate hypotheses, and connect knowledge across disparate areas. (Buehler 2023)

  • Consider using the Recurrent Memory Transformer (RMT) architecture to extend the context length of BERT, allowing for the storage and processing of both local and global information, and enabling information flow between segments of the input sequence through the use of recurrence. (A. Bulatov et al. 2023)

  • Consider using task-aware automatic prompt generation (TAP) to create high-quality prompts for multi-task pre-training, which can lead to significant improvements in model performance across various dialog-related tasks. (Y. Cai et al. 2023)

  • Explore integrating new technologies with Generative AI algorithms, such as reinforcement learning from human feedback (RLHF) and stable diffusion, to enhance the reliability, accuracy, and creativity of AI-generated content. (J. Cao et al. 2023)

  • Consider using a prompt-based path prediction method called DiscoPrompt to improve the performance of Implicit Discourse Relation Recognition (IDRR) tasks by incorporating the hierarchical structure of discourse relations and connectives into pre-trained language models. (C. Chan et al. 2023)

  • Treat evaluation as an essential discipline to better aid the development of large language models (LLMs), emphasizing the importance of understanding your strengths and weaknesses, guiding human-LLM interaction, ensuring safety and reliability, and adapting evaluation protocols to accommodate evolving LLM capabilities. (Y. Chang et al. 2023)

  • Consider using a combination of pre-trained text encoders, semantic tokenization using VQGAN, and masked generative transformers for efficient and effective text-to-image generation. (H. Chang et al. 2023)

  • Consider implementing speculative sampling, a novel algorithm that enables the acceleration of transformer decoding by generating multiple tokens from each transformer call, thereby reducing sampling latency without altering the target model or biased sample distribution. (Charlie Chen et al. 2023)

  • Consider implementing self-debugging methods in large language models to improve code generation performance, particularly through rubber duck debugging techniques that involve the model explaining its own code and identifying errors without requiring explicit human feedback. (Xinyun Chen et al. 2023)

  • Consider using a Soft Prompt-Based Calibration (SPeC) pipeline to reduce performance variability in clinical note summarization tasks while preserving the advantages of prompt-based summarization. (Y.-N. Chuang et al. 2023)

  • Utilise the softmax regression problem, which involves minimising the objective function (_{x{d}}|(Ax),{n}^{-1}(Ax)-b |{2}{2}), to improve the efficiency and effectiveness of your statistical analyses. (Yichuan Deng, Li, and Song 2023)

  • Employ bibliometric and discourse analyses to synthesize over 5,000 publications on large language models (LLMs) to identify research trends, patterns in research paradigms and collaborations, and the dynamic, fast-paced evolution of LLMs research. (L. Fan et al. 2023)

  • Develop a comprehensive evaluation benchmark for multimodal large language models (MLLMs) that assesses both perception and cognitive abilities across 14 subtasks, uses manually designed instruction-answer pairs to minimize data leakage, and enables quantitative analysis through concise instructions. (Chaoyou Fu et al. 2023)

  • Consider using an iterative algorithm for rescaled hyperbolic function regression when working with large language models, as it provides an input sparsity time algorithm applicable to functions like cosh() and sinh(), and can be adapted for in-context learning. (Yeqi Gao, Song, and Yin 2023)

  • Use hierarchy-aware attention when working with CLIP models to improve performance on vision and vision-language downstream tasks. (Geng et al. 2023)

  • Focus on understanding and mitigating the out-of-distribution factors contributing to length generalization failures in large language models, such as unseen distances, unseen number of tokens under attention, and implicitly encoded positional information, in order to improve your performance on longer text sequences. (C. Han et al. 2023)

  • Carefully consider the ethical implications of using ChatGPT in real-world applications, taking into account its potential to produce biased responses and perpetuate harmful language patterns, and employing effective prompt engineering techniques to mitigate these risks. (Hariri 2023)

  • Carefully consider the choice of GPT models, prompting strategies, and integration with traditional NMT systems when evaluating the performance of GPT models for machine translation, as these factors significantly influence translation quality. (Hendy et al. 2023)

  • Integrate domain knowledge into automated machine learning (AutoML) processes through Context-Aware Automated Feature Engineering (CAAFE), which leverages large language models to generate semantically meaningful features and explanations of your utility, thereby complementing existing automated feature engineering and AutoML methods. (Hollmann, Müller, and Hutter 2023)

  • Utilize a large pre-trained language model (LLM) to effectively extract API entities and relations from unstructured text, thereby reducing labor overhead and enhancing the overall process. (Qing Huang et al. 2023)

  • Consider the memory wall problem in generative LLM inference, where memory bandwidth is the primary bottleneck rather than compute, and therefore focus on optimizing memory usage through techniques like quantization. (Sehoon Kim et al. 2023)

  • Consider the unique challenges posed by reinforcement learning (RL) when attempting to apply transformer architectures, such as non-stationarity induced by changing policies, sensitivity to design choices, high computational and memory costs, and the need for more training data than models relying on strong inductive biases. (Youjia Li, Shi, and Zhang 2023)

  • Carefully examine the relationship between the attention mechanism and softmax unit in large language models, as it plays a crucial role in determining your performance across various natural language processing tasks. (Shuai Li et al. 2023)

  • Integrate aspect extraction and aspect-based recommendation tasks in an end-to-end manner, allowing for the generation of personalized and contextualized aspects that enhance the overall recommendation process. (Lei Li, Zhang, and Chen 2023)

  • Consider employing parameter-efficient fine-tuning (PEFT) methods when working with large language models, as these techniques allow for effective adaptation of the models without requiring extensive computational resources. (Lialin, Deshpande, and Rumshisky 2023)

  • Consider combining foundation models with millions of APIs to create a diverse and adaptable AI ecosystem that can handle both digital and physical tasks, improving interpretability and lifelong learning capabilities. (Yaobo Liang et al. 2023)

  • Carefully select appropriate benchmark datasets and evaluation metrics to assess the performance of chatbot systems in handling diverse Text-to-SQL tasks. (A. Liu et al. 2023)

  • Consider employing a fully-connected vision-language cross-modal connector in your Large Multimodal Models (LMMs), as it demonstrates remarkable power and data-efficiency, leading to superior performance across numerous benchmarks. (Haotian Liu, Li, Li, et al. 2023)

  • Explore the potential benefits of leveraging large language models like GPT-4 to generate multimodal instruction-following data for improving the performance of large multimodal models in handling diverse and challenging application-oriented tasks. (Haotian Liu, Li, Wu, et al. 2023)

  • Consider using ChatGPT as a general-purpose recommendation model, focusing on how its extensive linguistic and world knowledge acquired from large-scale corpora can be effectively transferred to recommendation scenarios, and evaluating its performance across various recommendation tasks. (Jiawei Liu et al. 2023)

  • Leverage the in-context learning capability of large language models (LLMs) by constructing dynamic contexts using domain-specific, individualized data, allowing the model to learn contextual knowledge from semantically similar examples from existing data. Furthermore, implementing an iterative optimization algorithm that performs automatic evaluation on the generated impression results and composes the corresponding instruction prompts can further optimize the model. (Chong Ma et al. 2023)

  • Consider using ChatGPT to filter and transform raw descriptions into high-quality captions for audio-language multimodal learning tasks, as demonstrated by the creation of the large-scale WavCaps dataset. (Mei et al. 2023)

  • Consider implementing gisting, a technique that compresses prompts into smaller sets of “gist” tokens, which can be cached and reused for computational efficiency, leading to significant improvements in processing speed and storage savings. (Mu, Li, and Goodman 2023)

  • Carefully consider the level of specificity in your prompts when working with large language models (LLMs), as it can significantly affect the accuracy, efficiency, and reliability of the generated code. (Murr, Grainger, and Gao 2023)

  • Utilize the REFINER framework, which involves a generator model producing intermediate reasoning steps and a critic model providing structured feedback on those steps, leading to improved reasoning performance in various natural language tasks. (D. Paul et al. 2023)

  • Utilize Task-Specific Prompts (TSP) and Domain-Specific Prompts (DSP) to optimize ChatGPTs performance in machine translation tasks, particularly in non-English-centric and cross-domain scenarios.’ (K. Peng et al. 2023)

  • Consider using soft prompts with frozen large language models (LLMs) for clinical concept and relation extraction, as it enables better few-shot learning and transfer learning capabilities, reduces computational costs, and allows for multi-task applications. (C. Peng et al. 2023)

  • Focus on developing efficient subquadratic primitives like element-wise multiplication and long convolutions to create larger convolutional language models, guided by mechanistic interpretability tasks such as recall and induction. (Poli et al. 2023)

  • Use a combination of natural language processing techniques, including transformer-based decoders and causal language modeling objectives, to effectively generate profile sentences from dialogues, thereby enabling the creation of more personalized and human-like conversational systems. (R. Ribeiro, Carvalho, and Coheur 2023)

  • Utilize the “Toolformer” model, which is capable of teaching itself to use external tools via simple APIs, thereby enabling it to achieve superior performance across a wide array of downstream tasks without compromising its core language modeling abilities. (Schick et al. 2023)

  • Consider using Decomposed Prompt Tuning (DePT) instead of traditional Prompt Tuning (PT) for parameter-efficient fine-tuning (PEFT) in order to save significant memory and time costs while maintaining or even improving performance across various natural language processing (NLP) and vision-language (VL) tasks. (Z. Shi and Lipani 2023)

  • Utilise the gem5 CPU simulator for evaluating performance, as it offers a completely deterministic assessment, guaranteeing both dependability and repeatability. (Shypula et al. 2023)

  • Focus on developing architectures like Retentive Network (RetNet) that enable simultaneous achievement of training parallelism, low-cost inference, and good performance for large language models. (Yutao Sun et al. 2023)

  • Carefully select appropriate natural language processing (NLP) tools based on your technical abilities and goals, and apply these tools to a diverse dataset of scientific texts to accurately evaluate sentiment and potential biases related to chronic Lyme disease. (Susnjak 2023)

  • Consider using a two-stage framework (ChatIE) to transform the zero-shot information extraction task into a multi-turn question-answering problem, leveraging the capabilities of large language models like ChatGPT to achieve impressive performance and potentially surpass traditional fully-supervised models. (Xiang Wei et al. 2023)

  • Consider the trade-offs between using unified large language models versus locally fine-tuned models for highly specific radiology NLI tasks, as the former may offer better performance with less data. (Zihao Wu et al. 2023)

  • Carefully evaluate the performance of large language models (LLMs) in translating natural language goals to structured planning languages, taking into account your sensitivity to prompts and limitations in handling numerical or physical reasoning tasks. (Y. Xie et al. 2023)

  • Leverage ChatGPT to automatically generate a high-quality multi-turn chat corpus, subsequently employing parameter-efficient tuning to enhance LLaMA, an open-source large language model, resulting in the creation of Baize, a highly capable open-source chat model. (Canwen Xu et al. 2023)

  • Consider using multiple prompt inputs instead of relying solely on selecting a better set of data samples within a single prompt input to optimize large language model (LLM) performance. (B. Yao et al. 2023)

  • Consider implementing a modularized training paradigm for large language models, combining a foundation LLM, a visual knowledge module, and a visual abstractor module, to enable effective multi-modal learning and improve overall performance. (Qinghao Ye et al. 2023)

  • Consider incorporating a planning algorithm alongside a pre-trained code generation Transformer to enhance the quality of generated programs. (Shun Zhang et al. 2023)

  • Carefully choose prompt examples for machine translation tasks, taking into account factors such as translation quality, semantic similarity, language model likelihood, sequence length, and similarity to test inputs, as these features can lead to improved translation performance. (Biao Zhang, Haddow, and Birch 2023)

  • Explore the scaling effect on model capacity by increasing the parameter scale of pre-trained language models (PLMs) to an even larger size, leading to the development of large language models (LLMs) with stronger capabilities in solving various natural language processing (NLP) tasks. (W. X. Zhao et al. 2023)

  • Consider utilizing large language models (LLMs) like ChatGPT for intelligent traffic safety research and applications, as they offer potential improvements in areas such as accident report automation, traffic data augmentation, and multisensory safety analysis. (O. Zheng et al. 2023)

  • Adopt a comprehensive approach to studying Pretrained Foundation Models (PFMs) by considering your historical evolution, underlying principles, diverse applications, and associated challenges in order to effectively leverage your potential for achieving artificial general intelligence. (Ce Zhou et al. 2023)

  • Focus on increasing the diversity and quality of your training data, rather than solely relying on larger quantities of data, to improve the performance of your language models. (Chunting Zhou et al. 2023)

  • Carefully consider the role of prompt engineering in maximizing the efficiency and accuracy of text generation using large language models (LLMs), while acknowledging the potential risks associated with misinformation and biases inherent in these models. (Teubner et al. 2023)

  • Carefully select appropriate clinical vignettes and compare your diagnostic accuracy against those generated by AI chatbots like ChatGPT-3 to understand the effectiveness of these tools in providing accurate differential diagnoses. (Hirosawa et al. 2023)

  • Consider the use of multi-modal pre-trained big models (MM-PTMs) for improved generalization and extraction of common features across multiple modalities, while being mindful of the challenges involved in acquiring and cleaning large-scale multi-modal data, designing appropriate network architectures and pre-training objectives, supporting large-scale computing power, and honing skills in parameter tuning. (Xiao Wang et al. 2023)

  • Carefully evaluate the human-likeness of synthetic data generated by large language models (LLMs) like GPT-3, as they show promise in providing valuable insights into human behavior, but may still exhibit biases and factual errors. (Hämäläinen, Tavast, and Kunnari 2023)

  • Consider using large language models (LLMs) for enabling conversational interaction with mobile UIs, as they demonstrate strong generalizability across multiple tasks and require minimal adaptation efforts. (Bryan Wang, Li, and Li 2023)

  • Evaluate the generalizability of existing misinformation detection models on AI-generated text, as they may not be as effective against this newer form of misinformation. (Jiawei Zhou et al. 2023)

  • Consider implementing a “PagedAttention” algorithm, which is inspired by the concept of paging in operating systems, to improve the efficiency of memory management in large language models (LLMs) and enhance overall performance. (W. Kwon et al. 2023)

  • Consider using a combination of PrOmpt Distillation (POD) and Task-alternated Training strategies to effectively and efficiently integrate multiple recommendation tasks into a Large Language Model (LLM). (Lei Li, Zhang, and Chen 2023)

  • Consider employing a combination of repetitiveness reduction techniques, skew alleviation strategies, and modeling heterogeneity approaches when dealing with highly repetitive, skew-distributed, and heterogeneous data in Ethereum transactions. (Sihao Hu et al. 2023)

  • Consider integrating knowledge-guided optimization in an iterative empirical framework to improve accessibility to research innovation through attainable resources. (Elnaggar et al. 2023)

  • Consider utilizing publicly available biomedical data for developing language models, as it may lead to similar or better performance compared to highly specialized private data collected from hospital reports or larger corpora with only generic data. (Labrak et al. 2023b)

  • Consider developing a VN-Transformer architecture to enhance the performance of your models in handling rotation-equivariant tasks, particularly in areas such as motion forecasting and 3D perception. (S. Khan et al. 2022)

  • Carefully consider the trade-offs between latency and throughput when selecting the best multi-dimensional partitioning technique for Transformer-based models, taking into account factors such as model size, sequence length, and available hardware resources. (Aminabadi et al. 2022)

  • Utilise large language models (LLMs) for efficient tabular data classification, particularly in situations where limited training data is available. (Carballo et al. 2022)

  • Consider using prompt tuning for speech processing tasks, as it allows them to optimize a limited number of task-specific parameters with a fixed pre-trained model, resulting in improved computation and memory efficiency. (K.-W. Chang et al. 2022)

  • Aim to create a unified foundation model for industrial recommender systems, capable of supporting open-ended domains and tasks through efficient adaptation, thereby reducing the need for extensive data collection and minimizing the carbon footprint associated with training separate models for each task. (Z. Cui et al. 2022)

  • Utilise the Polyglot Prompting (PolyPrompt) framework for multilingual multitask prompt training, which allows for the integration of different tasks from different languages into a monolithic framework without needing any task/language-specific modules. This approach was proven effective in improving performance across six tasks, covering 24 datasets and 49 languages. (J. Fu, Ng, and Liu 2022)

  • Utilise a pre-train, prompt and predict’ paradigm for solving type inference problems in programming languages. This involves pre-training a masked language model (MLM) on a large corpus of source code, then using a small amount of compiled library source code to stimulate the pre-trained MLM to recognise specific function qualifier names (FQNs) and your usage patterns. Afterwards, the type inference point in the partial code is converted into (Qing Huang et al. 2022)

  • Consider using prompt learning to effectively adapt your tasks to pre-trained language models, enabling direct modeling of text and leveraging the vast knowledge contained within these models. (Lei Li, Zhang, and Chen 2022)

  • Consider implementing a curriculum learning based prompt tuning (CUP) approach for implicit Event Argument Extraction (EAE), which involves parsing the document as an Abstract Meaning Representation (AMR) graph, followed by a series of four learning stages that increase in complexity, and finally utilising a prompt-based encoder-decoder model to elicit relevant information from pre-trained language models (PLMs) in each stage. (J. Lin et al. 2022)

  • Incorporate a memory-based feedback system into your language models to improve your ability to accurately interpret user intentions and reduce misunderstandings, thereby improving overall performance. (Madaan et al. 2022)

  • Aim to balance cooperation and specialization in multi-task learning systems through the use of a modularized mixture of experts (Mod-Squad) model, which optimizes the joint distribution over tasks and experts to encourage a sparse but strong dependence between them. (B. Mustafa et al. 2022)

  • Explore the relationship between in-context learning in Transformers and gradient-based meta-learning formulations, particularly focusing on the potential equivalence between data transformations induced by a single linear self-attention layer and those resulting from gradient-descent on a regression loss. (Oswald et al. 2022)

  • Focus on optimizing the design and training procedures of vision transformers, specifically through parallelizing the architecture, fine-tuning attention layers, and incorporating patch pre-processing with masked self-supervised learning. (Touvron et al. 2022)

  • Focus on developing more flexible and heterogeneous transformer architectures, rather than simply compressing existing ones, in order to achieve significant improvements in performance. (Tuli et al. 2022)

  • Utilise CP-Tuning, an end-to-end Contrastive Prompt Tuning framework for fine-tuning Pre-trained Language Models without requiring any manual engineering of task-specific prompts and verbalizers. (Ziyun Xu et al. 2022)

  • Consider using a systematic approach for prompt design in relation extraction tasks for a specific domain, including a variety of ranking scores for prompt selection, as it can improve model performance in both fine-tuned and few-shot training conditions. (H.-S. Yeh, Lavergne, and Zweigenbaum 2022)

  • Consider learning pluggable, interpretable, and extensible prompts to enhance pre-trained protein models by injecting task-specific knowledge, leading to improved performance in protein function and structure prediction tasks. (Qiang Zhang et al. 2022)

  • Utilize Transformer architectures, specifically TransTEE, for treatment effect estimation because it demonstrates superior performance compared to other methods while being highly efficient and versatile in handling various types of treatments and covariates. (Y.-F. Zhang et al. 2022)

  • Employ generative pre-trained transformers (GPT) to automate the early-stage design concept generation process, enabling the transformation of knowledge and reasoning from textual data into new concepts in understandable language. (Qihao Zhu and Luo 2022)

  • Utilize a generative pre-trained language model (PLM) to automatically retrieve and map biological analogy and generate bio-inspired design (BID) concepts in the form of natural language. (Qihao Zhu, Zhang, and Luo 2022)

  • Utilize long-sequence transformer models, such as Longformer and BigBird, to effectively analyze long clinical texts and capture long-term dependencies, leading to improved performance across various clinical NLP tasks. (Yikuan Li et al. 2022)

  • Consider using a unified “Pretrain, Personalized Prompt, and Predict Paradigm” (P5) for recommendation tasks, which involves pretraining a model on a large-scale personalized prompt collection covering various recommendation task families, allowing it to understand unseen personalized prompts and generalize to novel personalized prompts or unseen items in other domains. (Geng et al. 2022)

  • Consider using a unified CRS approach based on knowledge-enhanced prompt learning, specifically the UniCRS model, to address the issue of semantic inconsistency between recommendation and conversation modules in conversational recommendation systems. (Xiaolei Wang et al. 2022)

  • Consider utilizing a text-to-text Transformer language model for generating political event data from unstructured text, overcoming the limitations of dictionary methods and classifier-based approaches. (Yaoyao Dai, Radford, and Halterman 2022)

  • Consider utilising pre-trained image and text transformer models for optical character recognition (OCR) tasks, as demonstrated by the proposed TrOCR model, which shows improved performance over traditional CNN-based methods. (Atienza 2021)

  • Consider developing unified end-to-end models for video-language pre-training, such as the All-in-One Transformer’, which effectively integrates raw video and textual signals into joint representations using a unified backbone architecture. This approach allows for improved efficiency and state-of-the-art performance on various downstream video-text tasks.’ (Bertasius, Wang, and Torresani 2021)

  • Consider using a unified framework like Lavender for video-language understanding tasks, as it simplifies model architecture, enables seamless support for all downstream tasks with a single set of parameter values, generalizes well to various downstream tasks with limited training samples, and allows for zero-shot evaluation on video question answering tasks. (Bogolin et al. 2021)

  • Consider adopting a parallel design strategy when integrating MobileNet and transformer architectures, utilizing a two-way bridge to enable bidirectional fusion of local and global features. (Yinpeng Chen et al. 2021)

  • Consider employing a unified approach when dealing with complex multimodal data, specifically combining vision, text, and layout modalities, to improve overall performance and efficiency in document artificial intelligence tasks. (Xingyu Chen et al. 2021)

  • Consider utilizing factorized reshaped matrices when working with natural language processing models, as they offer significant reductions in the number of trainable parameters while preserving model performance. (Fedus, Zoph, and Shazeer 2021)

  • Focus on developing large, diverse datasets for training deep neural networks, rather than spending excessive effort on designing custom network architectures. (Hawthorne et al. 2021)

  • Consider using LoRA (Low-Rank Adaptation) for large-scale pre-training and adaptation to particular tasks or domains, as it significantly reduces the number of trainable parameters and GPU memory requirements, while maintaining or improving model quality. (E. J. Hu et al. 2021)

  • Leverage a BERT-based siamese architecture for real-time document ranking in web search engines, as it significantly improves production performance while maintaining efficiency. (Kocián et al. 2021)

  • Focus on developing and testing methods that enable the reuse of a single frozen model for multiple downstream tasks, as this can lead to significant efficiency gains and improved performance. (Lester, Al-Rfou, and Constant 2021)

  • Consider using prompt-based few-shot learning techniques for conversational AI tasks, as they offer comparable results to fully trained models while being more computationally efficient and easier to maintain. (Madotto et al. 2021)

  • Utilize a unified framework called LFPT5, which employs prompt tuning of T5 to enable lifelong few-shot language learning across diverse tasks and domains, thereby reducing overfitting and catastrophic forgetting. (Chengwei Qin and Joty 2021)

  • Focus on creating a generative model that incorporates long-range coherence within the pretraining data, allowing the model to infer a shared latent concept across examples, thus facilitating in-context learning. (S. M. Xie et al. 2021)

  • Consider developing a data mining pipeline to collect a large-scale rap dataset with aligned rhythmic beats, and subsequently design a Transformer-based autoregressive language model that carefully models rhymes and rhythms within the context of rap generation. (L. Xue et al. 2021)

  • Consider combining convolutional neural networks (CNNs) with transformers to create a hybrid structure that combines your respective strengths in handling local and global information for pixel-wise prediction tasks. (Guanglei Yang et al. 2021)

  • Utilise the BertRL (BERT-based Relational Learning) model for relation prediction in knowledge graphs, as it outperforms current state-of-the-art methods in both inductive and transductive settings, demonstrates strong generalisation capabilities in few-shot learning, and offers explainability. (Zha, Chen, and Yan 2021)

  • Consider using the proposed Attention Free Transformer (AFT) model, which eliminates the need for dot product self-attention and thus reduces memory complexity, leading to improved efficiency and competitive performance compared to traditional Transformers and other variants. (Zhai et al. 2021)

  • Consider using the Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners (DART) technique when working with pre-trained language models for few-shot learning tasks, as it allows for improved performance through differentiable template and label optimization. (N. Zhang et al. 2021)

  • Consider using a multimodal transformer-based pre-training model like MEmoBERT, combined with a prompt-based learning approach, to improve the efficiency and effectiveness of your multimodal emotion recognition tasks. (Jinming Zhao et al. 2021)

  • Focus on developing a two-step framework for improving the efficiency and effectiveness of pre-trained language models (PLMs) in online ranking systems, involving the extraction of query-dependent summaries and the implementation of a modularized PLM like Pyramid-ERNIE, which decouples text representation and interaction to strike a balance between efficiency and effectiveness. (L. Zou et al. 2021)

  • Focus on developing a diversity-driven ensemble of convolutional autoencoders (CAE-Ensemble) for accurate and efficient outlier detection in time series data, incorporating a diversity metric, parameter transfer-based training strategy, and unsupervised hyperparameter selection. (Campos et al. 2021)

  • Focus on developing efficient and stable training methods for large-scale transformer models, such as ViT-22B, to achieve superior performance in various computer vision tasks. (Abnar and Zuidema 2020)

  • Consider using curriculum learning when developing open-domain chatbots, as it allows for gradual learning of response generation, moving from simple one-to-one mappings to more complex one-to-many mappings, ultimately improving overall performance. (S. Bao et al. 2020)

  • Utilize citation graphs as a powerful signal of document-level relatedness to improve the quality of document-level embeddings generated by transformer language models. (Cohan et al. 2020)

  • Consider running multiple trials with different random seeds to achieve substantial gains in performance on various datasets, particularly when dealing with smaller datasets. (Dodge et al. 2020)

  • Focus on developing dynamic BERT models that offer flexibility in both width and depth directions, enabling a richer range of architectural configurations and better exploration of the balance between model accuracy and size. (L. Hou et al. 2020)

  • Consider using transformer models for non-recurrent handwritten text-line recognition, as these models offer improved accuracy and efficiency compared to traditional recurrent neural networks. (L. Kang et al. 2020)

  • Consider incorporating adversarial training into your BERT language model fine-tuning processes for aspect extraction and aspect sentiment classification tasks in sentiment analysis, as it has been demonstrated to significantly improve model performance. (Karimi, Rossi, and Prati 2020)

  • Consider using contextual embeddings, such as ELMo and BERT, in your natural language processing tasks because these models go beyond global word representations and provide context-dependent representations that can capture many syntactic and semantic properties of words under diverse linguistic contexts. (Qi Liu, Kusner, and Blunsom 2020)

  • Consider implementing time-restricted self-attention for the encoder and triggered attention for the encoder-decoder attention mechanism when developing a transformer-based end-to-end ASR system for streaming ASR tasks. (Moritz, Hori, and Roux 2020)

  • Utilise the Mixed Interest Network (MiNet) model to accurately predict Click Through Rates (CTR) in cross-domain scenarios by considering three types of user interests: long-term interest across domains, short-term interest from the source domain, and short-term interest in the target domain. (Ouyang et al. 2020)

  • Consider utilising AdapterHub, a framework that simplifies the process of training and sharing adapters for transformer-based models like BERT, RoBERTa, and XLM-R, thereby improving efficiency, reducing storage requirements, and promoting modularity and composition of information across various tasks and languages. (Pfeiffer et al. 2020)

  • Consider employing a two-stage search strategy when dealing with long sequential user behavior data, allowing for more accurate and efficient modelling of user interests. (Pi Qi et al. 2020)

  • Consider implementing the Emformer approach for low latency streaming speech recognition, as it offers improved efficiency and reduced computational complexity compared to other existing approaches like AM-TRF. (Yangyang Shi et al. 2020)

  • Apply a data balance training strategy to improve the performance of your multilingual text-to-speech model, particularly for low-resource languages. (Jingzhou Yang and He 2020)

  • Utilise the SLED (Sliding-Encoder and Decoder) technique for efficiently processing long text sequences, which involves breaking down the input into overlapping chunks, encoding each with a short-text LM encoder, and then using the pretrained decoder to combine information across chunks. (Ainslie et al. 2020)

  • Use a novel model called “AprilE” which employs triple-level self-attention and pseudo residual connection to model relational patterns, particularly symmetric and antisymmetric relations, in knowledge graph embedding tasks. (Yuzhang Liu et al. 2020)

  • Consider utilizing a Multi-Modal Transformer-based (MMTrans) approach when dealing with code summarization tasks for smart contracts. This approach effectively addresses the challenge of extracting semantic information from source code by leveraging two modalities of the Abstract Syntax Tree (AST): Structure-based Traversal (SBT) sequences and graphs. By doing so, the MMTrans can capture both global and local semantic information, enabling it to generate higher-quality code comments compared to existing (LeClair et al. 2020)

  • Consider incorporating mixed interest network (MiNet) models in your study designs, which utilizes two levels of attention mechanisms to accurately predict click-through rates (CTR) in cross-domain scenarios by considering long-term interests across domains, short-term interests from the source domain, and short-term interests in the target domain. (Ouyang et al. 2020)

  • Consider utilizing pre-trained models (PTMs) for natural language processing (NLP) tasks, as they offer numerous benefits such as learning universal language representations, providing better model initializations, acting as a form of regularization against overfitting, and being able to be fine-tuned for specific downstream tasks. (XiPeng Qiu et al. 2020)

  • Adopt a unified text-to-text transformer framework for natural language processing tasks, allowing them to convert all text-based language problems into a text-to-text format, thereby enabling comparisons of pre-training objectives, architectures, unlabelled datasets, transfer approaches, and other factors across numerous language understanding tasks. (Anil et al. 2019)

  • Consider using light-weight adapter layers in neural machine translation models to achieve simultaneous adaptation to multiple individual tasks, thereby improving efficiency and scalability. (Bapna, Arivazhagan, and Firat 2019)

  • Consider developing generative models of commonsense knowledge to improve automatic knowledge base construction, as demonstrated by the success of COMET in producing high-quality, diverse commonsense descriptions in natural language. (Bosselut et al. 2019)

  • Consider using a multi-stage fusion transformer to effectively capture long-range dependencies within neural network architectures, enabling accurate prediction of attributes such as latency and accuracy. (Han Cai et al. 2019)

  • Carefully consider factors such as latency, scale, personalization, fairness and privacy, and metrics design when developing large-scale language models for email assistance tools like Smart Compose. (M. X. Chen et al. 2019)

  • Utilize the Transformer model to effectively capture sequential signals within users behavior sequences for improved recommendation system performance.’ (Q. Chen et al. 2019)

  • Utilize extensive data augmentation and initialization with pre-trained weights to optimize the performance of your Transformer-based encoder-decoder model for automatic speech recognition (ASR). (Hrinchuk, Popova, and Ginsburg 2019)

  • Consider implementing parameter-reduction techniques like factorized embedding parameterization and cross-layer parameter sharing to enhance the scalability and efficiency of pre-trained language models, leading to improved performance on various downstream tasks. (Z. Lan et al. 2019)

  • Leverage pre-trained language models like BERT for improved performance in emotion classification tasks, particularly in cases where labeled data is limited. (L. Luo and Wang 2019)

  • Consider combining text representations with metadata and knowledge graph embeddings to enhance the performance of deep neural language models like BERT for document classification tasks. (Ostendorff et al. 2019)

  • Consider utilising pre-trained checkpoints for sequence generation tasks, as demonstrated through the development of a Transformer-based sequence-to-sequence model that is compatible with publicly available pre-trained BERT, GPT-2, and RoBERTa checkpoints. This approach resulted in new state-of-the-art results on Machine Translation, Text Summarization, Sentence Splitting, and Sentence Fusion. (Rothe, Narayan, and Severyn 2019)

  • Employ a novel approach called G-BERT, which integrates Graph Neural Networks (GNNs) and BERT (Bidirectional Encoder Representations from Transformers) for medical code representation and medication recommendation, thereby overcoming selection bias and lack of hierarchical knowledge in traditional approaches. (Shang et al. 2019)

  • Utilize layer-wise relevance propagation (LRP) to identify the most important heads in each encoder layer of a transformer model, followed by characterizing the roles performed by these heads, and finally employing a pruning method based on stochastic gates and a differentiable relaxation of the (L_{0}) penalty to remove redundant heads without significantly affecting performance. (Voita et al. 2019)

  • Consider the contextual nature of entities and relations in knowledge graphs, rather than assigning a single static representation to each entity/relation, and utilize transformer encoders to generate contextualized representations for improved link prediction and path query answering. (Quan Wang et al. 2019)

  • Consider combining self-attention and recurrent neural network structures in your transformer models to enhance decoding efficiency and maintain translation quality. (Chengyi Wang, Wu, and Liu 2019)

  • Leverage the power of Transformer architectures and pretraining techniques to improve the performance of natural language processing models, while utilizing the open-source Transformers’ library to simplify the process of implementing, distributing, and adapting these advanced models.’ (Wolf et al. 2019)

  • Consider adopting a hybrid text normalization system that combines the strengths of rule-based models and neural models, particularly for Mandarin text normalization tasks where previous studies often rely solely on hand-crafted rules. (Jinjin Zhang et al. 2019)

  • Utilise a layer-wise visualisation technique to gain insights into the internal workings of Transformer networks, allowing them to identify incorrect predictions and determine which aspects of the context were deemed significant by the model. (Aken et al. 2019)

  • Focus on improving the efficiency of Transformer models by addressing the quadratic complexity issue of the self-attention mechanism and reducing computation costs through techniques such as pooling and sparsity. (Voita et al. 2019)

  • Utilize a combination of pre-training and supervised fine-tuning in order to maximize the flexibility and adaptability of your models. (Al-Rfou et al. 2018)

  • Consider leveraging synthetic instructions as an intermediate representation to bridge the gap between human language and agent understanding, allowing for improved generalization and easier learning. (P. Anderson et al. 2018)

  • Consider using a discriminative model for pre-training text encoders, as opposed to a generative model, since it allows the model to learn from all input tokens rather than just a small masked-out subset, resulting in improved computational efficiency and superior downstream performance. (Caccia et al. 2018)

  • Consider decomposing the decoding process into two stages when working with semantic parsing, allowing them to better handle complex meaning representations. (L. Dong and Lapata 2018)

  • Utilize a combination of techniques including careful parameter initialization, denoising effects of language models, and iterative back-translation to develop successful unsupervised machine translation models. (Lample et al. 2018)

  • Consider using automated methods to generate diverse and high-quality prompts for querying language models, rather than relying solely on manually crafted prompts, in order to more accurately estimate the knowledge contained within the models. (McCann et al. 2018)

  • Consider integrating multiple types of auxiliary information, such as geographic and social attributes, spatial dependencies, and online crowd queries, using a hybrid Seq2Seq model for improved accuracy in traffic prediction. (B. Liao et al. 2018)

  • Consider implementing the ProphetNet model, which utilizes a novel self-supervised objective called “future n-gram prediction” and the proposed n-stream self-attention mechanism, to improve the efficiency and accuracy of sequence-to-sequence pre-training tasks. (X. Du, Shao, and Cardie 2017)

  • Consider utilizing a multimodal sequence to sequence learning architecture to improve the accuracy and quantity of API mappings during the process of migrating APIs between different programming languages. (X. Gu et al. 2017)

  • Incorporate event circumstances into narrative event prediction using the CircEvent approach, which utilizes multi-head attention mechanisms to capture both local and global event circumstances. (Zhouhan Lin et al. 2017)

  • Consider incorporating a double attention block into your deep neural networks to improve efficiency and accuracy in image and video recognition tasks by allowing the network to aggregate and propagate informative global features from the entire spatio-temporal space of input images/videos. (L.-C. Chen et al. 2016)

  • Consider employing a bi-directional attention flow model for machine comprehension tasks, as it significantly outperforms existing methods on the Stanford Question Answering Dataset (SQuAD) test set leaderboard and achieves state-of-the-art results on the CNN/DailyMail cloze test. (Seo, Kembhavi, et al. 2016)

  • Carefully choose the appropriate pre-training objective and fine-tuning strategy for your specific problem, taking into account factors such as model complexity, data availability, and desired level of performance. (J. L. Ba, Kiros, and Hinton 2016)

  • Consider developing general purpose vision systems capable of learning and performing various tasks without modifying the architecture or learning process, thereby reducing development time and increasing versatility. (Yonghui Wu et al. 2016)

  • Consider using MetaL-Prompt, a novel lightweight automatic prompt generation method for Language-Model-as-a-Service (LMaaS), which meta-trains a prompt generation model (PGM) to enable robust learning by the language model from the contexts created by the generated prompts, allowing the PGM to generate prompts for unseen tasks without requiring additional training for those specific tasks, and significantly reducing computational costs compared to previous methods. (Xiang Zhang, Zhao, and LeCun 2015)

  • Consider implementing hierarchical memory networks (HMNs) for large-scale factoid question answering tasks, as they offer a scalable and efficient alternative to traditional soft and hard attention mechanisms. (Auvolat et al. 2015)

  • Consider using a semantic structure-based approach to predict query graphs from natural language questions, which can help filter out noisy candidate query graphs and improve overall accuracy in complex question answering over knowledge graphs. (Bordes et al. 2015)

  • Carefully consider the potential effects of punctuation on the performance of natural language processing models, particularly in tasks like natural language inference, and ensure that your experimental designs adequately address this issue. (Bowman, Angeli, et al. 2015)

  • Consider utilising the sequence to sequence (seq2seq) model for conversational modelling tasks, as it enables end-to-end training and reduces reliance on hand-crafted rules. (Vinyals and Le 2015)

  • Utilise a two-stage learning approach for TinyBERT, incorporating both general distillation and task-specific distillation, to ensure accurate knowledge transfer from the larger teacher’ BERT model.’ (Yunchao Gong et al. 2014)

  • Combine a position-aware attention mechanism with an LSTM sequence model to enhance relation extraction performance, and utilize a large, supervised dataset like TACRED to train the model effectively. (Zaremba, Sutskever, and Vinyals 2014)

  • Consider replacing traditional Hidden Markov Models (HMM) in continuous speech recognition tasks with a bi-directional recurrent neural network encoder coupled to a recurrent neural network decoder that utilizes an attention mechanism to establish alignment between input and output sequences, resulting in improved phoneme error rates compared to existing methods. (Chorowski et al. 2014)

  • Consider using a distribution rectification distillation (DRD) technique to mitigate the impact of query information distortion in low-bit quantized DETR (Q-DETR) systems, thereby enabling improved performance and reduced computational requirements. (Yoshua Bengio, Léonard, and Courville 2013)

  • Consider developing a closed-loop speech chain model based on deep learning to integrate human speech perception and production behaviors, allowing for better understanding and improvement of speech processing systems. (Sakti et al. 2012)

  • Utilise the Birds Eye probe, a novel information-theoretic approach, to effectively detect whether and how contextualised text representations encode linguistic graph structures. (Recht and Re 2012)

  • Consider utilizing self-supervised pretraining for protein modeling tasks, as it shows promise in improving performance across various applications, despite requiring further development for optimal results. (NA?)

  • Carefully choose appropriate table serialization methods and context inclusion strategies depending on the target downstream task, as these factors significantly affect the performance of transformer-based models for tabular data representation. (NA?)

  • Carefully consider factors such as latency, scale, personalization, fairness and privacy, and metrics design when developing large-scale language models for email assistance tools like Smart Compose. (NA?)

  • Utilise a deep-meta-learning approach when dealing with complex spatiotemporal correlations in urban traffic prediction. This involves creating a model called ST-MetaNet that uses a sequence-to-sequence architecture, incorporating an encoder to learn historical information and a decoder to make predictions step by step. The encoder and decoder have identical network structures, featuring a recurrent neural network to encode traffic, a meta graph attention network to capture diverse spatial correlations, and a (NA?)

  • Consider utilising pre-trained checkpoints for sequence generation tasks, as demonstrated through the development of a Transformer-based sequence-to-sequence model that achieves state-of-the-art results on multiple benchmarks while significantly reducing computational costs. (NA?)

  • Employ a combination of tensor and sequence parallelism to optimize memory usage and minimize redundancy in large transformer models, thereby improving overall performance. (NA?)

  • Consider domain-specific pretraining from scratch for specialized domains like biomedicine, as it can lead to significant improvements in performance compared to continually pretraining general-domain language models. (NA?)

  • Carefully evaluate and select appropriate semantic models for your specific NLP tasks, considering factors such as language complexity, resource availability, and task requirements, and potentially developing custom models tailored to your needs. (NA?)

  • Integrate Self-Supervised Attention (SSA) into BERT models to improve your generalization capabilities and reduce overfitting on smaller datasets. (NA?)

  • Utilize deep learning methods to enable semantic communication systems, specifically focusing on joint semantic-channel coding, to improve overall system capacity and reduce semantic errors in text transmission. (NA?)

  • Employ a deep learning architecture called “Enformer” to predict gene expression and chromatin states from DNA sequences, as it effectively integrates information from long-range interactions (up to 100 kb away) in the genome, leading to more accurate variant effect predictions on gene expression for both natural genetic variants and saturation mutagenesis measured by massively parallel reporter assays. (NA?)

  • Utilise the ART (Autoencoding-based Retriever Training) technique for unsupervised learning of dense retrieval models, which involves generating soft labels for passage retrieval based on question reconstruction probabilities. (NA?)

  • Consider using a combination of content-based features and contextual information to effectively characterize malware propagation and evolution patterns for improved malware detection. (NA?)

  • Utilize a combination of natural language processing techniques, such as rule-based and machine learning approaches, along with the BERT model for emotion classification, to effectively analyze sentiment in social media data regarding the coronavirus pandemic. (NA?)

  • Utilize a scoring system to select appropriate dialogue samples for few-shot training in dialogue summarization tasks, leading to improved performance in terms of ROUGE scores and human evaluations. (NA?)

  • Consider employing deep learning techniques for gap-filling in time series data, as demonstrated through the successful implementation of a deep learning method for accurately filling gaps in eddy covariance crop evapotranspiration data. (NA?)

  • Consider implementing the URRBP (Ugly Requests Require Beautiful Prompts) approach to enhance AI accessibility and interaction, which includes a human-optimized layer for prompt optimization and an intelligent system for AI model selection, thereby addressing the complexities of prompt engineering and diverse AI model interactions. (NA?)

  • Consider adopting prompt tuning as an efficient and effective method for fine-tuning protein language models for specific tasks, while acknowledging the importance of developing robust evaluation metrics that accurately reflect the biological relevance of the generated sequences. (NA?)

  • Consider the rapid development and deployment of advanced AI chatbots, such as ChatGPT, in higher education settings, and evaluate your potential impact on academic integrity, pedagogical practices, and student learning experiences. (NA?)

  • Be aware of the ethical implications of using AI in academic writing, and should strive to maintain the integrity and credibility of the scientific community by adhering to established publication ethics guidelines. (NA?)

  • Develop a novel Chinese few-shot text classification method called CIPLUD, combining an improved prompt learning method and existing unlabeled data for Chinese FSTC, to improve classification performance. (NA?)

  • Consider employing a knowledge-guided prompt learning method for few-shot text classification, which involves revealing relevant knowledge for text classification through a combination of a knowledge prompting template and two multi-task frameworks. (NA?)

  • Consider using a contrastive sample method based on knowledge-guided prompt learning (ConKgPrompt) for text classification, which leverages external knowledge bases to expand the label vocabulary and incorporates supervised contrastive learning to make representations more expressive. (NA?)

  • Carefully evaluate the accuracy and reliability of AI-generated content before incorporating it into your work, especially in fields such as nuclear medicine where patient safety and healthcare quality depend on precise and reliable information. (NA?)

  • Consider adopting a virtual prompt pre-training method for prototype-based few-shot relation extraction, which projects the virtual prompt to latent space and fuses with PLM parameters, thereby improving the interaction with the PLM and avoiding laborious and subjective label word mapping and prompt template engineering. (NA?)

  • Consider employing a cascade prompt learning framework with a sentence-level attention mechanism (CasATT) for discriminative question answering, which involves mining evidence accurately from large-scale documents by retrieval and ranking, and answering questions with ranked candidates. (NA?)

  • Consider employing a few-shot learning paradigm to evaluate systematic generalization in human subjects, allowing them to assess the ability to learn the meaning of words from limited examples and generalize to more complex instructions. (NA?)

  • Consider adopting the Regression Transformer (RT) methodology, which allows for concurrent sequence regression and generation for molecular language modelling, leading to improved performance in property prediction and conditional sequence generation tasks. (NA?)

  • Consider utilizing Transformer-based language models, specifically TransPolymer, for accurate and efficient polymer property predictions due to its ability to effectively learn intrinsic relationships between polymers and your properties, reduce overfitting, and achieve state-of-the-art results on various benchmarks. (NA?)

  • Carefully evaluate the quality and reliability of AI-generated scientific abstracts before incorporating them into your work, considering factors like plagiarism detection, AI output detection, and adherence to journal formatting standards. (NA?)

  • Consider utilizing deep transfer learning techniques, specifically BERT-NLI, to improve the efficiency and effectiveness of your supervised machine learning models, especially when dealing with scarce or imbalanced data. (NA?)

  • Adopt a language-specialised data-centric approach, combining transfer learning techniques and language-specific subword methods, to effectively capture the unique morphological and syntactic structures of the target language, leading to improved translation performance, particularly for low-resource languages. (NA?)

  • Carefully consider the type of prompt template when applying prompt learning for news recommendation tasks, as different templates can lead to varying levels of success depending on factors like semantic relevance, user emotion, user action, and recommendation utility. (NA?)

  • Consider employing prompt-based learning (PBL) alongside fine-tuning with pre-trained language models (PLMs) like BERT to effectively classify temporal relationships between treatments and hospitalisation times in clinical texts, achieving high levels of accuracy and efficiency. (NA?)

  • Consider three primary dimensions when evaluating large language models: what to evaluate, where to evaluate, and how to evaluate. (NA?)

  • Carefully consider and address issues of interpretability, experimental bias, and practical applicability when developing and evaluating automated program generation models. (NA?)

Graph Neural Networks (Gnn)

  • Carefully choose appropriate graph-based deep learning methods for fake news detection, taking into account the specific type of graph structure and the nature of the data available. (S. Gong et al. 2023)

  • Consider using Neural meta-Graph Search (NGS) for explainable graph neural network-based fraud detection, which involves formalizing the message passing process of GNN using meta-graph, searching the meta-graph using differentiable neural architecture search (DARTS), and aggregating node embeddings captured by multiple meta-graphs. (Zidi Qin et al. 2022)

  • Aim to optimize your models by identifying and eliminating unnecessary computations, thereby increasing efficiency without sacrificing accuracy. (Hisadome and Matsui 2021)

  • Consider integrating adversarial learning into the boosting training procedure of Gradient Boosting Decision Trees (GBDT) to improve unsupervised domain adaptation for malware detection, while also proposing a new instance weighting scheme to reduce the negative impact of incorrect pseudo labels. (Panpan Qi et al. 2021)

  • Use a combination of label-balanced sampling and neighborhood sampling strategies to overcome the class imbalance issue in graph-based fraud detection tasks. (Yang Liu et al. 2021)

  • Adopt a two-tower architecture for deep neural networks when dealing with candidate retrieval tasks, specifically focusing on a multi-head design for the query tower and an attention-based loss function for improved semantic understanding and efficiency. (Han Zhang et al. 2020)

  • Consider both sequence information and spatial information when developing models for charge prediction, as demonstrated by the successful implementation of the SECaps model. (C. He et al. 2019)

  • Focus on developing unsupervised methods for detecting online fraud reviewer groups, rather than relying solely on supervised approaches like frequent itemset mining, because unsupervised methods can better handle complexities such as unclear definitions of groups, variations in inter-group dynamics, and scarcity of labeled group-level spam data. (Dhawan et al. 2019)

  • Adopt a Bayesian framework to induce the hypothesis space when dealing with unsupervised domain adaptation tasks, optimizing the embedding and kernel of the Gaussian process so that the posterior hypothesis distribution leads to consistent class predictions, reducing the maximum classifier discrepancy. (Minyoung Kim et al. 2019)

  • Consider using a multi-scale 3D domain adaption network called PointDAN when working with point cloud data, as it enables simultaneous alignment of global and local features, leading to improved accuracy in tasks such as classification, detection, and segmentation. (Can Qin et al. 2019)

  • Employ a two-stage approach when dealing with complex time-evolving graph problems. First, they should use a Long Short-Term Memory R-GCN (LRGCN) to capture both temporal dependency and structure dynamics. Afterwards, they should implement a self-attentive path embedding (SAPE) technique to convert paths of varying lengths into fixed-length vectors, thereby enabling better classification performance and providing meaningful interpretation of the underlying data. (Jia Li et al. 2019)

  • Consider using graph neural networks (GNNs) to analyze complex relationships in datasets, particularly where there may be concept drift, label uncertainty, or excessive human effort involved in traditional fraud detection methods. (C. Liang et al. 2019)

  • Develop a two-stage recommender system consisting of a matching phase followed by a ranking phase, with the matching phase focusing on computing pairwise similarities between items based on user behavior, and the ranking phase leveraging deep neural networks to rank candidate items based on user preferences. (Jizhe Wang et al. 2018)

  • Utilise the E2PN model for efficient SE(3)-equivariant feature learning from 3D point clouds, which combines group convolutions and quotient representations to reduce computational complexity and memory consumption, while preserving the capacity to distinguish rotations. (T. Cohen et al. 2017)

  • Explore unsupervised domain adaptation using adversarial neural networks to train a segmentation method that is more invariant to differences in input data, thereby improving the performance of automatic segmentation systems on new data that differs from the training data. (Kamnitsas et al. 2016)

  • Focus on developing methods that can effectively extract meaningful subgraphs from larger graphs, allowing for easier understanding and analysis of complex data structures. (Serban, Lowe, et al. 2016)

  • Consider extending the CORAL algorithm to learn a nonlinear transformation that aligns correlations of layer activations in deep neural networks, resulting in improved performance in unsupervised domain adaptation tasks. (B. Sun and Saenko 2016)

  • Consider using deep subspace clustering methods that operate outside of the traditional self-expressive framework, allowing for linear time and space complexity, scalability to large datasets, and applicability to online clustering scenarios. (Diederik P. Kingma and Ba 2014)

  • Consider incorporating hyperbolic metric learning and hierarchical clustering techniques when working with complex, non-Euclidean data structures, as they can help to effectively excavate richer similarity information beyond binary in modeling. (Diederik P. Kingma and Ba 2014)

  • Consider implementing a Multi-Stage Self-Supervised (M3S) Training Algorithm for Graph Convolutional Networks (GCNs) to enhance generalization performance on graphs with few labeled nodes, particularly through the use of a multi-stage training framework and DeepCluster technique. (Bruna et al. 2013)

Meta-Learning

  • Use objective Bayesian inference procedures for the parameters of the multivariate random effects model, specifically employing the Berger and Bernardo reference prior and the Jeffreys prior, to ensure accurate estimation of the overall mean vector and the between-study covariance matrix. (Bodnar and Bodnar 2023)

  • Carefully consider the task landscape when implementing Model-Agnostic Meta-Learning (MAML), particularly focusing on the hardness and geographical distribution of tasks, as these factors greatly influence the effectiveness of MAML over Non-Adaptive Learning (NAL). (Collins, Mokhtari, and Shakkottai 2020)

  • Adopt a meta-learning approach in order to improve the learning algorithm itself, allowing for enhanced data and computational efficiency, as well as better generalization capabilities. (Hospedales et al. 2020)

  • Utilize knowledge distillation to develop efficient student models based on MobileNetV3, applying a combination of novel architectural modifications and existing speed-up techniques such as low-rank matrix approximation and weight quantization to optimize student embeddings for mobile devices. (Shor et al. 2020)

  • Develop a meta-optimizer that learns in the space of both point-based and population-based optimization algorithms, while balancing exploration and exploitation through the inclusion of a posterior and entropy term in the meta-loss function. (K. Cao et al. 2019)

  • Utilize a bi-level optimization based on meta-learning to directly optimize the network for few-shot class incremental learning, thereby aligning the training objectives with the actual evaluation goals. (Clune 2019)

  • Carefully evaluate the suitability of traditional AutoML tools for data stream mining settings, taking into account factors like concept drift and the availability of the entire training dataset. (Elshawi, Maher, and Sakr 2019)

  • Utilize deep meta-reinforcement learning (meta-RL) for Neural Architecture Search (NAS) to efficiently adapt previously learned policies rather than starting from scratch, thereby significantly reducing computational costs while maintaining high performance. (Robles and Vanschoren 2019)

  • Aim to bridge the gap between artificial intelligence and human cognition by developing machine learning techniques that enable computers to learn from a limited number of examples, thereby reducing the need for large-scale datasets and enabling rapid generalization. (Yaqing Wang et al. 2019)

  • Adopt a hierarchically structured meta-learning (HSML) algorithm to efficiently handle task uncertainty and heterogeneity in meta-learning, by explicitly tailoring transferable knowledge to different clusters of tasks, thus improving overall performance. (Huaxiu Yao et al. 2019)

  • Carefully examine the role of feature reuse versus rapid learning in your meta-learning algorithms, as the authors demonstrate that feature reuse plays a dominant role in the success of Model Agnostic Meta-Learning (MAML) and propose the ANIL (Almost No Inner Loop) algorithm as a computationally beneficial alternative. (Antoniou, Edwards, and Storkey 2018)

  • Consider integrating unsupervised meta-learning with simple task construction mechanisms, such as clustering embeddings, to achieve improved performance on various downstream, human-specified tasks. (K. Hsu, Levine, and Finn 2018)

  • Utilize a Dirichlet process mixture of hierarchical Bayesian models over the parameters of an arbitrary parametric model like a neural network to handle latent distribution shifts in meta-learning and continual learning situations. (Jerfel et al. 2018)

  • Aim to meta-learn an unsupervised learning rule that directly targets later desired tasks, rather than simply minimizing a surrogate objective like negative log likelihood of a generative model. (Metz et al. 2018)

  • Consider incorporating unlabelled data and distractor classes into your few-shot learning scenarios to create a more realistic and challenging environment for evaluating the effectiveness of your models. (M. Ren et al. 2018)

  • Adopt a systematic and standardized approach to computing meta-features for classification datasets in order to improve the reproducibility and comparability of meta-learning studies. (Rivolli et al. 2018)

  • Consider incorporating metric scaling and task conditioning when developing few-shot learning algorithms, as these techniques can significantly enhance the performance of such models. (Bauer et al. 2017)

  • Carefully consider the evolution of metalearning concepts, particularly regarding the characterization of the metalearning process, the nature of meta-knowledge, extending algorithm selection to automatic design of solutions, and understanding the range of potential application domains. (Brazdil and Giraud-Carrier 2017)

  • Utilize meta-learning strategies to efficiently optimize machine learning models by leveraging prior experiences and meta-data from similar tasks, thereby reducing the need for extensive experimentation and manual parameter tuning. (Lorena et al. 2017)

  • Leverage the power of meta-learning by adjusting priors based on the Extended PAC-Bayes theory, enabling efficient learning of novel future tasks by effectively capturing the common structure across previously learned tasks while retaining adequate flexibility to adapt to unique aspects of new tasks. (Amit and Meir 2017)

  • Consider using gradient-based meta-learning techniques, such as model-agnostic meta-learning (MAML), due to your high representational power and improved statistical efficiency compared to traditional recurrent models. (Finn and Levine 2017)

  • Develop a meta optimizer to mitigate catastrophic forgetting in sequential domain meta-learning (SDML) by dynamically adjusting learning rates for meta parameters, balancing between remembering previous domains and efficiently learning the current domain. (Vinyals et al. 2016)

  • Employ inverted regularization at the inner loop and ordinary regularization at the outer loop during training to improve the generalization capabilities of your meta-models. (Alexander A. Alemi et al. 2016)

  • Avoid the “memorization problem” in meta-learning by carefully designing your meta-training tasks to ensure they are mutually exclusive, thereby forcing the model to adapt to new tasks rather than relying solely on previously learned information. (Alexander A. Alemi et al. 2016)

  • Consider utilizing a meta-learning framework for solving cold-start recommendation problems, especially in scenarios where new items arrive continuously, as it offers flexibility in combining user and item information, enables the use of deep neural networks for non-linear embeddings, and facilitates efficient transfer learning across users. (Hidasi et al. 2015)

  • Consider exploring meta-learning the mean function of a Gaussian process prior, as it can be useful in the meta-learning setting and reduce the risk of overfitting compared to standard supervised learning. (Rusk 2015)

  • Consider using latent embedding optimization (LEO) for meta-learning tasks, as it enables efficient adaptation in low-dimensional latent spaces while still being capable of generating high-dimensional model parameters. (V. Mnih et al. 2015)

  • Adopt a meta-learning approach to develop a regularizer that enables domain generalization, allowing models to better handle domain shifts and improve overall performance. (K. He et al. 2015a)

  • Focus on defining and transferring high-level distilled knowledge as the flow for solving a problem, rather than just mimicking intermediate results, to achieve faster optimization, improved performance in small networks, and effective transfer learning. (Branson et al. 2014)

  • Consider using memory-augmented neural networks (MANNs) for meta-learning tasks, as they have demonstrated the capability to rapidly assimilate new data and make accurate predictions after only a few samples, while also introducing a new method for accessing external memory that focuses on memory content rather than location-based focusing mechanisms. (Graves, Wayne, and Danihelka 2014)

  • Compare the performance of stacking methods for combining classifiers to the performance of simply selecting the best classifier from the ensemble through cross-validation, as the former may not always provide superior results. (NA?)

  • Consider implementing intrinsic motivation systems in autonomous mental development, specifically through the use of intelligent adaptive curiosity (IAC) algorithms, which promote active exploration and learning by focusing on maintaining maximum learning progress in novel situations. (NA?)

  • Carefully consider the prerequisites and potential pitfalls of implementing metalearning systems, such as the representativeness of extracted metafeatures, the novelty of the problem domain, and the reliability of performance estimates, in order to ensure effective and reliable use of metalearning in various machine learning applications. (NA?)

  • Move beyond focusing solely on learning algorithms and instead consider the broader implications of lifelong machine learning (LML) systems, which involve the retention and application of learned knowledge over time. (NA?)

  • Utilise the A3R metric, which combines accuracy and runtime, in your algorithm selection processes to improve mean interval loss values. (NA?)

Supervised Learning Algorithms

Linear Regression

  • Utilise penalised regression techniques in selecting high-dimensional control variates to gain performance benefits over traditional least squares methods. (South et al. 2023)

  • Utilise orthogonal subsampling (OSS) for big data, which involves seeking subsamples with maximum combinatorial orthogonality to improve efficiency and accuracy in linear regression modelling. (Lin Wang et al. 2021)

  • Use a BIC-type penalty for optimal aggregation in regression models, as it achieves simultaneously the (L), (C), and (MS) bounds of the form (1.4) with the optimal rates Δn,M = ψn,M. (NA?)

  • Utilise the adaptive lasso technique over the traditional lasso approach in order to achieve both consistent variable selection and optimal prediction in your statistical modelling. (NA?)

Logistic Regression

  • Utilize a trust region Newton method for large-scale logistic regression, as it provides fast convergence and outperforms other common approaches like quasi-Newton methods. (NA?)

  • Utilize a trust region Newton method for large-scale logistic regression, as it provides fast convergence and outperforms other common approaches like quasi-Newton methods. (NA?)

Decision Trees

  • Consider adopting an interactive machine learning (IML) model instead of traditional classical machine learning (CML) methods, particularly when dealing with complex datasets involving numerous features and requiring rapid feedback cycles for accurate classification. (Wondimu, Buche, and Visser 2022)

  • Carefully consider various classification algorithms such as decision trees, rule-based methods, nearest neighbor methods, support vector machines, neural networks, ensemble methods, and others when working with data streams to ensure accurate and efficient results. (Peipei Li et al. 2022)

  • Focus on developing a geometric scoring method for ranking cases within decision trees, which preserves the intelligibility of the model while providing an effective ranking mechanism. (Hustad et al. 2021)

  • Utilise a combination of sparse relaxation and argmin differentiation to effectively learn binary trees, allowing them to simultaneously learn the continuous parameters of splitting decisions and provide a principled approach to learning tree pruning. (Zantedeschi, Kusner, and Niculae 2020)

  • Carefully address the issue of absent levels in decision tree algorithms, as they can cause biased results and negatively impact model performance. (Au 2017)

  • Consider implementing the Bonsai algorithm for efficient prediction in resource-constrained Internet of Things (IoT) environments, as it offers high prediction accuracy while fitting within tight memory budgets. (Ioannou et al. 2017)

  • Consider extending Random Forests (RFs) to the one-class setting by developing a natural methodology to adapt standard splitting criteria, allowing for more effective anomaly detection and one-class classification. (Goix et al. 2016)

  • Utilize a novel algorithm for generating optimal sparse decision trees, which combines analytical bounds to reduce the search space and modern systems techniques like data structures and a custom bit-vector library, resulting in improved scalability, speed, and proof of optimality. (Gunluk et al. 2016)

  • Utilise decision trees for building models in online advertising contexts, as they effectively handle categorical features and offer representation discovery’, allowing for adaptation to changing environments. (Kalyanakrishnan, Singh, and Kant 2014)

  • Consider implementing an online incremental algorithm to compute the heuristic measure of an attribute with reduced computational cost, followed by selecting a subset of attributes to identify potential split timings, ultimately leading to significant reductions in computational costs and split-delays while enhancing overall model accuracy. (Sovdat 2014)

  • Utilise the Multivariate GUIDE methodology for analysing multiple response variables, as it provides a robust framework for unbiased variable selection and accurate predictions. (Loh and Zheng 2013)

  • Consider implementing a non-trivial enumeration algorithm for identifying all distinct decision trees within a given feature subset, allowing for efficient pruning of redundant trees and significant improvements in computational performance. (Aldinucci, Ruggieri, and Torquati 2013)

  • Utilise decision trees to determine the optimal sequence of arguments for persuading individuals, taking into consideration the potential variations in individual beliefs and preferences. (Hunter 2013)

  • Investigate the potential benefits of incorporating subtree replacement (also known as grafting) alongside traditional pruning methods in decision tree simplification processes, as it may lead to statistically significant reductions in tree size without compromising accuracy. (Ruggieri 2012)

  • Utilize a combination of visualization, interaction, and algorithmic support to effectively create and analyze decision trees while incorporating domain-specific knowledge. (Elzen and Wijk 2011)

  • Consider using the “node harvest” technique when working with complex datasets, as it allows for greater flexibility in determining the appropriate weights for different nodes within the dataset, leading to potentially better predictive accuracy. (Meinshausen 2010)

  • Integrate discrimination awareness directly into the model induction process of a decision tree, specifically through dependency-aware tree construction and leaf relabeling, to minimize discrimination while maintaining high predictive accuracy. (Kamiran, Calders, and Pechenizkiy 2010)

  • Focus on developing parsimonious and highly predictive tree models through controlling the search for local interactions, employing efficient variable and split selection strategies, and fitting nontrivial models in the nodes. (Loh 2009)

  • Consider using multivariate dyadic regression trees (MDRTs) for sparse learning problems, as they offer simultaneous adaptation to the unknown sparsity and smoothness of the true regression functions while achieving near-optimal rates of convergence. (Blanchard et al. 2007)

  • Focus on developing scalable algorithms for decision tree construction while maintaining the quality of the tree, as demonstrated by the RainForest framework. (“Advanced Data Mining and Applications” 2005)

  • Utilize more powerful classification strategies at tree leaves in your incremental tree induction methods, such as naive Bayes classifiers, to significantly enhance the overall performance of your decision models. (Gama, Rocha, and Medas 2003)

  • Push size and accuracy constraints into the tree-building phase of decision tree construction, rather than applying them as an afterthought, to achieve significant performance improvements. (Garofalakis et al. 2000)

  • Utilise association rules for classification purposes by building a decision tree-like structure using those rules and leveraging the accuracy-driven pruning of decision tree induction. (Ke Wang, Zhou, and He 2000)

  • Consider utilizing decision trees for failure diagnosis in complex internet systems, as demonstrated by the authors successful identification of 13 out of 14 true causes of failure in eBay’s system.’ (NA?)

  • Focus on developing decision trees for classification tasks, using a top-down induction approach, while considering the trade-off between model complexity and interpretability. (NA?)

  • Consider converting continuous attributes into ordered discrete attributes prior to feeding them into a learning system, as this can lead to significant improvements in learning time without compromising accuracy. (NA?)

  • Utilize the information entropy minimization heuristic for discretizing continuous-valued attributes in decision tree generation, as it provides a better understanding of the heuristic, offers formal justification for its usage, and improves computational efficiency. (NA?)

  • Consider the tradeoff between model complexity and interpretability when selecting a decision tree algorithm, as oversimplified trees may lack sufficient detail to accurately represent complex relationships within the dataset, while overcomplicated trees may become difficult to interpret and potentially less effective in predicting future instances. (NA?)

  • Focus on developing decision tree algorithms that minimize the sum of misclassification and test costs, through the use of a novel splitting criterion for attribute selection and intelligent test strategies that can suggest ways of obtaining missing values at a cost. (NA?)

  • Leverage machine learning techniques, specifically decision trees, to conduct decision point analysis in the context of process mining, allowing them to understand how data attributes influence the choices made in a business process. (NA?)

  • Consider utilizing decision trees for classification tasks due to your interpretability, flexibility, and ability to handle mixed feature types; however, care must be taken to prevent overfitting through techniques like stopping splits early or pruning the tree based on cross-validated error rates. (NA?)

  • Focus on estimating the VC dimension of decision trees using the partitioning function, which can lead to improved performance in pruning algorithms compared to traditional methods like CART. (NA?)

  • Adopt a decision tree framework for spatiotemporal sequence prediction, which involves decomposing the prediction task into a series of overlapping fixed-length multivariate regression problems that can be effectively handled by decision trees. (NA?)

  • Focus on using multiple parameters, specifically the size or depth of the decision tree, maximum domain size of all features, and the maximum Hamming distance between any two examples, to ensure fixed-parameter tractability in decision tree learning. (NA?)

  • Utilise mixed-integer optimization (MIO) techniques to build optimal decision trees for your datasets, as this approach offers superior accuracy and interpretability compared to traditional heuristic methods. (NA?)

  • Consider implementing the RapidScorer algorithm for tree ensemble evaluation, which uses a modified run length encoding called epitome to optimize memory usage and improve traversal speed, resulting in significant improvements in computational efficiency compared to existing methods. (NA?)

Random Forests

  • Utilize the ranger’ software for high dimensional data analysis due to its superior scalability, runtime efficiency, and memory optimization when compared to alternative random forest implementations.’ (Wright and Ziegler 2017)

  • Consider using boosted decision tables instead of boosted regression trees due to your superior accuracy and efficiency in terms of scoring latency. (Y. Lou and Obukhov 2017)

  • Carefully choose the subsampling rate and tree depth parameters in random forests models to optimize your performance. (Duroux and Scornet 2016)

  • Utilise Causal Forest, a non-parametric method for heterogeneous treatment effect estimation, which yields valid asymptotic confidence intervals for the true underlying treatment effect, thereby providing a robust solution for analysing treatment effect heterogeneity. (Wager and Athey 2015)

  • Utilise the redundancies inherent within the label space to increase the efficiency of your classification processes., ‘The main methodological insight provided by this paper is the recommendation to leverage the redundancies found within the label space to enhance the effectiveness of classification procedures.’ (R. Yan, Tesic, and Smith 2007)

  • Avoid using the Gini importance measure for variable selection in random forests, especially when dealing with datasets containing mixed-type predictor variables, and instead opt for the permutation importance measure combined with subsampling without replacement. (C. Strobl et al. 2006)

  • Consider using the Extra-Trees algorithm, which involves selecting splits, both attribute and cut-point, either completely or partially at random, to achieve improved accuracy and computational efficiency in supervised classification and regression problems. (NA?)

  • Avoid using traditional random forest variable importance measures when dealing with datasets containing predictor variables that vary significantly in scale levels or category numbers, due to the inherent bias in those measures. Instead, they suggest utilizing an alternative random forest algorithm that offers unbiased variable selection in individual classification trees, and applying it through subsampling without replacement to obtain accurate and reliable variable importance measures. (NA?)

  • Use conditional permutation importance measures rather than marginal importance measures when dealing with correlated predictor variables in random forests, as this helps to accurately capture the true impact of individual variables on the response. (NA?)

  • Consider combining the strengths of random forests Gini importance for feature selection with regularized linear classifiers, such as discriminant partial least squares regression, to achieve optimal results in analyzing high-dimensional spectral data.’ (NA?)

  • Consider using permutation importance (PIMP) to correct for biased measures of feature importance in machine learning models, thereby improving model interpretability and prediction accuracy. (NA?)

  • Consider using random forest classifiers with repeated random sub-sampling to effectively manage highly imbalanced data in disease prediction models. (NA?)

  • Use random forest models to analyze the relationship between various predictor variables and extinction risk in marine mammals, taking into account both intrinsic and extrinsic factors, while considering the limitations of available data. (NA?)

  • Utilize a separate model for each hourly period, employing component-wise gradient boosting to estimate each model using univariate penalized regression splines as base learners, allowing for the electricity demand to change with time-of-year, day-of-week, time-of-day, and on public holidays, with the main predictors being current and past temperatures as well as past demand. (NA?)

  • Consider utilizing quantum principal component analysis (QPCA) for analyzing unknown density matrices, as QPCA offers an exponential speed-up over traditional algorithms. (NA?)

  • Consider utilizing Random Forest (RF) machine learning technique instead of multiple linear regression (MLR) for predicting crop yields, as RF demonstrated superior performance across all tested scenarios. (NA?)

  • Utilize the Recursive Feature Elimination (RFE) algorithm for variable selection in high-dimensional regression or classification frameworks, specifically when dealing with correlated predictors. (NA?)

  • Utilize the iterative Random Forest algorithm (iRF) when dealing with high-dimensional genomics data to efficiently discover high-order interactions while maintaining computational feasibility. (NA?)

  • Consider applying the Random Forest (RF) algorithm to neuroimaging data for improved accuracy in distinguishing between stable MCI (sMCI) and progressive MCI (pMCI) that converts to Alzheimers disease (AD), due to its inherent feature selection capabilities and robustness to noise.’ (NA?)

  • Conduct large-scale benchmarking experiments to compare the performance of different machine learning algorithms, taking inspiration from clinical trial methodologies to minimize biases and ensure rigorousness. (NA?)

  • Consider alternative approaches to Random Forest-Recursive Feature Elimination (RF-RFE) when dealing with high-dimensional datasets containing numerous correlated variables, as RF-RFE might decrease the importance of causal variables along with correlated ones, making detection harder. (NA?)

Support Vector Machines (Svm)

  • Focus on establishing lower bounds on the fraction of support vectors in your models, particularly when working with universal kernels, to improve the performance and interpretability of your classifiers. (Haas et al. 2023)

  • Consider using localized multiple kernel learning (LMKL), which allows for the extraction of local importance of kernels, as opposed to traditional methods like mixture of experts or mixture of SVMs, which only provide global importance information. (Gautam et al. 2018)

  • Combine Kernel Canonical Correlation Analysis (KCCA) and Support Vector Machines (SVM) into a single optimization process called SVM-2K to improve the performance of classification tasks in high-dimensional feature spaces. (X. Xie and Sun 2015)

  • Utilize a cutting plane algorithm combined with multiple kernel learning to efficiently convert the (l_{0})-norm Sparse SVM (SSVM) into a mixed integer programming (MIP) problem, allowing for improved sparsity and generalization performance in high-dimensional datasets. (M. Tan, Tsang, and Wang 2013)

  • Adopt a Squared Loss Support Vector Machine (L2-SVM) approach with separate LASSO constraints over pre-treatment and causal heterogeneity parameters to effectively estimate heterogeneous treatment effects in complex scenarios involving numerous treatments and covariates. (Imai and Ratkovic 2013)

  • Consider implementing a randomized block-coordinate variant of the classic Frank-Wolfe algorithm for convex optimization with block-separable constraints, as it achieves a similar convergence rate in duality gap as the full Frank-Wolfe algorithm, while reducing computational costs. (Lacoste-Julien et al. 2012)

  • Consider the computational complexity of learning when selecting machine learning algorithms, as well as your performance in handling large datasets. (H. S. Chang, Weiss, and Freeman 2009)

  • Consider training Laplacian Support Vector Machines (LapSVMs) in the primal rather than the dual formulation, as it offers significant improvements in efficiency and reduces training time. (Melacci and Belkin 2009)

  • Combine both data balancing techniques and classifier modifications to effectively handle imbalanced data sets, specifically through the use of support vector machines with soft margins and boosting algorithms. (B. X. Wang and Japkowicz 2009)

  • Consider robustness as a fundamental property of classification algorithms, and utilize robustness arguments instead of traditional VC dimension or stability measures to demonstrate the consistency of support vector machines. (Huan Xu, Caramanis, and Mannor 2008)

  • Carefully select and evaluate the choice of convex loss function in your estimation scheme, as it significantly affects the accuracy and consistency of the resulting classifier. (Tong Zhang 2004)

  • Utilize the principles of Statistical Learning Theory, specifically Support Vector Machines, to effectively analyze and predict the location of devices within a Wi-Fi network based on Received Signal Strength Intensity (RSSI) measurements. (NA?)

  • Utilise coordinate descent methods when dealing with large linear SVM problems, as it provides an efficient and stable solution compared to other existing algorithms. (NA?)

  • Consider using robust optimization techniques in conjunction with regularization to improve the generalizability of your machine learning models, particularly in cases where the training data may be affected by non-iid disturbances. (NA?)

  • Utilise Support Vector Machines (SVMs) for pattern recognition tasks due to your ability to minimize structural risk, condense information into support vectors, and effectively operate in high-dimensional spaces. (NA?)

  • Use an online recursive algorithm for training support vector machines, one vector at a time, to efficiently handle large datasets and accurately assess generalization performance. (NA?)

  • Focus on developing methods that ensure uniform one-sided convergence of empirical risk to actual risk, as defined by the Key Theorem, in order to achieve optimal performance in statistical learning tasks. (NA?)

  • Focus on developing stable learning algorithms, which can lead to better generalization error bounds through the use of concentration inequalities. (NA?)

  • Utilise the powerful theorem due to Marshall and Olkin (1960) in conjunction with convex optimization techniques by Popescu and Bertsimas (2001) to provide bounds on the probability of misclassifying a point, without assuming any specific distribution. (NA?)

  • Utilise the maximum margin clustering technique for improved accuracy in your clustering tasks, particularly when dealing with nonlinear datasets. (NA?)

  • Utilise Support Vector Machines (SVMs) for tasks like chunk identification due to your ability to achieve high generalisation performance even when dealing with a large number of features. (NA?)

  • Consider utilizing the support vector machine (SVM) method for protein secondary structure prediction, as it demonstrates strong performance in terms of segment overlap measure (SOV) and three-state overall per-residue accuracy (Q3), while also offering benefits such as effective avoidance of overfitting and the ability to handle large feature spaces. (NA?)

  • Utilise a support vector machine (SVM) approach for the detection of microcalcifications in digital mammography, as it demonstrates superior performance compared to traditional methods. (NA?)

  • Leverage clickthrough data, which is readily available and relatively inexpensive, to train search engine algorithms using support vector machines (SVMs) within a risk minimization framework, ultimately improving retrieval quality. (NA?)

  • Utilise the Positive Definite Fuzzy Classifier (PDFC) when dealing with high dimensional feature spaces due to its capacity to maintain good generalisation abilities while avoiding the curse of dimensionality’. (NA?)

  • Consider using the Clustering-Based Support Vector Machine (CB-SVM) method for handling large datasets, as it employs a hierarchical micro-clustering algorithm to provide high-quality samples to the SVM, thereby improving both scalability and classification accuracy. (NA?)

  • Carefully evaluate various multicategory classification methods, including multicategory support vector machines (MC-SVMs), gene selection techniques, and cross-validation designs, to ensure the development of a robust and accurate cancer diagnostic model based on microarray data. (NA?)

  • Consider employing advanced machine learning techniques like support vector machines (SVM) alongside traditional statistical methods like backpropagation neural networks (BNN) to achieve improved prediction accuracy and enhanced interpretability in credit rating analysis across diverse markets. (NA?)

  • Consider utilising Support Vector Machines (SVMs) for classification tasks involving high-dimensional data, particularly in hyperspectral remote sensing applications, due to your ability to effectively manage the curse of dimensionality’ without requiring a feature selection step to reduce data dimensionality.’ (NA?)

  • Carefully select relevant features when building SVM models for load forecasting, and consider the potential impact of time-series concepts on improving forecast accuracy. (NA?)

  • Utilize the multicategory support vector machine (MSVM) for handling multicategory classification problems, as it effectively extends the binary SVM to the multicategory case and maintains the optimal property of the binary case, providing a unifying framework for both equal and unequal misclassification costs. (NA?)

  • Consider utilizing kernel-based learning methods, specifically those implemented in the kernlab package, due to your ability to implicitly map input data into higher dimensional feature spaces, allowing for effective classification, regression, and clustering tasks without explicit feature extraction. (NA?)

  • Consider using Least Squares Support Vector Machine (LS-SVM) classifiers for your studies because they offer comparable test set performances to traditional SVM classifiers, especially when combined with standard cross-validation procedures for hyperparameter selection. (NA?)

  • Utilize a multi-task learning approach based on the minimization of regularization functionals, allowing them to model the relationship between tasks using a novel kernel function that incorporates a task-coupling parameter. (NA?)

  • Use Support Vector Machines (SVMs) for classification and regression tasks, as they embody the Structural Risk Minimization (SRM) principle, which is superior to the traditional Empirical Risk Minimization (ERM) principle employed by conventional neural networks, leading to better generalization capabilities. (NA?)

  • Consider using multiple kernel learning (MKL) for improved interpretability and efficiency in large-scale optimization tasks, as demonstrated through the successful integration of MKL in the machine learning toolbox SHOGUN. (NA?)

  • Utilise a generalized version of multiclass support vector machines (SVMs) to effectively handle complex output structures in machine learning tasks. (NA?)

  • Consider employing data integration methods, such as median rank scores or quantile discretization, alongside established machine learning techniques like support vector machines, to improve the generalizability and reliability of predictive models in cross-platform classification analysis of gene expression data from different studies. (NA?)

  • Consider using Support Vector Machines (SVM) instead of traditional linear regression models when dealing with non-linear relationships in financial markets, due to SVMs ability to effectively capture complex patterns while avoiding overfitting.’ (NA?)

  • Consider utilizing one-class support vector machines (SVMs) instead of generating pseudo-absence data when dealing with presence-only datasets, as it allows for accurate predictions without altering the potential distribution area. (NA?)

  • Balance the k classes and implement a novel Newton refinement modification to PSVM in order to address the issue of unbalanced classes and achieve significant improvements in test set accuracy without sacrificing the speed of PSVM. (NA?)

  • Utilize combined SVM-based feature selection and classification methods to optimize your models, particularly when dealing with complex datasets. (NA?)

  • Consider employing high-dimensional non-linear pattern classification techniques to improve the accuracy of distinguishing truthful from non-truthful responses in neuroimaging studies. (NA?)

  • Focus on developing learning algorithms that directly optimize the desired performance metric rather than relying solely on traditional methods that optimize error rates. (NA?)

  • Consider utilizing Support Vector Machines (SVMs) for addressing complex classification tasks in bioinformatics, such as cancer diagnosis and protein secondary structure prediction, due to your ability to handle high-dimensional datasets with limited training samples effectively. (NA?)

  • Utilise Support Vector Machines (SVMs) to predict whether a new phenotype derived from a non-synonymous coding Single Nucleotide Polymorphism (nsSNP) can be related to a genetic disease in humans, achieving over 74% accuracy in doing so. (NA?)

  • Avoid using cross-validation (CV) to estimate the error of a classifier that has already been tuned using CV, because this leads to a significantly biased estimate of the true error. Instead, they recommend employing a nested CV procedure, which provides an almost unbiased estimate of the true error. (NA?)

  • Utilize a combination of machine learning algorithms to accurately predict intrachain bridges from sequence alone, solving the prediction problem in two steps - first predicting the disulfide bonding state of each cysteine by a binary classifier, and then pairing cysteines that are known to participate in the formation of bridges to obtain a connectivity pattern. (NA?)

  • Consider using the Cutting-Plane Algorithm when dealing with high-dimensional sparse data in machine learning applications, as it offers significant improvements in computational efficiency compared to traditional decomposition methods. (NA?)

  • Utilise Support Vector Machines (SVMs) for classification tasks, specifically because they reduce the classification problem to the computation of a linear decision function, do not suffer from local minima in the optimisation problem, and offer a computationally efficient decision function. (NA?)

  • Focus on improving the alignment between the kernel and the target function in order to enhance the performance of kernel-based learning algorithms. (NA?)

  • Consider using multiple kernel learning (MKL) for multiclass problems, as it enables the optimization of kernel weights while training the SVM, potentially improving classification accuracy and helping identify relevant features. (NA?)

  • Utilise a support vector method for optimising average precision in your studies, as it provides a globally optimal solution to a straightforward relaxation of MAP, improving the efficiency and effectiveness of your analyses. (NA?)

  • Focus on developing universally consistent classifiers that minimize classification risk, but recognize that achieving specific convergence rates for all distributions is impossible; instead, they should target smaller classes of distributions with appropriate assumptions about smoothness or decision boundaries. (NA?)

  • Consider using the hybrid huberized support vector machine (HHSVM) for microarray classification and gene selection because it overcomes the limitations of traditional SVM models by combining the huberized hinge loss function and the elastic-net penalty, allowing for improved variable selection results particularly when variables are highly correlated. (NA?)

  • Carefully consider the choice of feature selection methods, the number of genes in the gene list, and the number of cases (samples) when analyzing microarray gene expression data, as these factors significantly impact classification success. (NA?)

  • Leverage the parallel processing capabilities of graphics processing units (GPUs) to significantly speed up the training and classification processes of support vector machines (SVMs), resulting in substantial computational efficiency improvements. (NA?)

  • Utilize a dual coordinate descent method for large-scale linear support vector machines (SVM) because it provides an epsilon-accurate solution in O(log(1/epsilon)) iterations, making it significantly faster than existing state-of-the-art solvers. (NA?)

  • Consider the inverse relationship between the runtime of SVM optimization and the size of the training dataset, particularly when employing a subgradient descent approach like PEGASOS, as it may lead to improved efficiency and performance. (NA?)

  • Consider implementing the Granular Support Vector Machines - Repetitive Undersampling (GSVM-RU) algorithm for handling highly imbalanced classification tasks, as it effectively minimizes information loss during the undersampling process while simultaneously increasing the efficiency of the SVM prediction. (NA?)

  • Consider extending your structural SVM models to include latent variables, providing an efficient algorithm for solving the optimization problem of the proposed formulation, and applying this new algorithm to relevant problems such as discriminative motif finding in yeast DNA. (NA?)

  • Utilise the dual version of Ridge Regression combined with ANOVA enhanced infinite-node splines to effectively handle the curse of dimensionality’, thereby improving the accuracy of non-linear regression predictions.’ (NA?)

  • Carefully evaluate the choice of machine learning algorithms for hippocampal segmentation, considering factors such as accuracy, power to map disease effects, and dependence on the size of the training set, as demonstrated by comparing AdaBoost, Support Vector Machines, and other methods in the context of Alzheimers disease detection.’ (NA?)

  • Employ multiple validation methods such as leave-one-out cross-validation and train-and-test when validating your data mining frameworks, especially when dealing with limited training samples. (NA?)

  • Consider employing support vector machines (SVMs) for the analysis of microarray expression data, as they have demonstrated effectiveness in accurately classifying tissue samples and identifying potential errors within the data. (NA?)

  • Incorporate privileged information provided by an intelligent teacher during the training phase of machine learning models to improve the efficiency and accuracy of student learning. (NA?)

  • Employ a fast Newton method for selecting features in support vector machine classification, as it effectively handles high-dimensional spaces and large numbers of data points without requiring specialized linear programming packages, ultimately leading to more accurate and efficient classifications. (NA?)

  • Consider employing balanced learning with optimized decision making to improve the classification accuracy of microcalcification clusters in mammogram imaging, particularly when dealing with imbalanced datasets. (NA?)

  • Utilise the Rank-Loss Support Instance Machine (SIM) methodology for improving the accuracy of instance annotation tasks in Multi-Instance Multi-Label (MIML) settings. (NA?)

  • Carefully evaluate whether your data meets the criteria for “big data” before applying complex machine learning techniques, as simpler methods might suffice. (NA?)

  • Consider using a combination of feature selection and classification techniques, specifically support vector machines, to effectively analyze high-dimensional class-imbalanced datasets. (NA?)

  • Develop “safe” semi-supervised learning approaches, specifically S4VMs, which leverage multiple candidate low-density separators rather than relying solely on one optimal low-density separator, thereby reducing the risk of identifying a poor separator with unlabeled data and ensuring that your performance is never significantly inferior to inductive SVMs. (NA?)

  • Consider using a kernel matrix correction to enhance the robustness of support vector machines (SVMs) against adversarial data manipulation in classification tasks. (NA?)

  • Carefully select appropriate convex loss functions for your specific application contexts, as these functions play a crucial role in determining the accuracy and effectiveness of classification models. (NA?)

  • Carefully choose appropriate classifiers and preprocess your fMRI data effectively, while avoiding potential pitfalls such as peeking and violating independence assumptions, to accurately analyze and interpret your findings. (NA?)

  • Consider implementing a novel post-processing strategy based on calculating a new bias for Support Vector Machines (SVMs) to improve your performance on imbalanced datasets, without requiring modifications to the standard optimization problem or introduction of new parameters. (NA?)

  • Use active learning techniques with support vector machines to efficiently identify the most informative training examples, thereby increasing performance while minimizing costs associated with labeling. (NA?)

  • Use the SwissADME web tool to efficiently compute key physicochemical, pharmacokinetic, drug-like, and related parameters for one or multiple molecules, leveraging its open-access, fast, statistically significant, and intuitive predictive models. (NA?)

  • Focus on developing novel methods for selecting optimal subsets of training data for Support Vector Machines (SVMs) to overcome the challenge of high time and memory training complexities associated with large datasets. (NA?)

  • Consider utilizing Recursive Feature Elimination (RFE) algorithms for non-linear kernels in conjunction with Support Vector Machines (SVM) to effectively rank and visualize the importance of variables in biomedical studies, thereby enhancing understanding of mechanisms of association and reducing costs related to biomarker development. (NA?)

  • Focus on establishing lower bounds on the fraction of support vectors in your models, particularly when working with universal kernels like the Gaussian RBF kernel, to improve the performance and interpretability of your classifiers. (NA?)

  • Utilise the SVMTorch algorithm when dealing with large-scale regression problems due to its efficiency in solving such problems compared to traditional methods. (NA?)

  • Utilize a multi-task learning approach based on the minimization of regularization functionals, allowing them to model the relationship between tasks using a novel kernel function that incorporates a task-coupling parameter. (NA?)

Naïve Bayes Classifier

  • Consider employing a selective Bayesian classifier algorithm to improve the accuracy of your models in domains with correlated features, while preserving the benefits of the simpler naive Bayesian classifier in domains without such correlations. (Langley and Sage 2013)

  • Consider the possibility of the Bayesian classifier being optimal even when the independence assumption is violated, as it can outperform more complex models in certain scenarios. (Domingos and Pazzani 1997)

K-Nearest Neighbors (K-Nn)

  • Consider applying different termination conditions for each query in approximate nearest neighbor (ANN) search problems, rather than relying solely on static features such as the query vector itself. By incorporating runtime features such as intermediate search results, researchers can build prediction models that achieve the same accuracy with less overall search effort compared to fixed configurations. (Conglong Li et al. 2020)

  • Consider using approximate nearest neighbor search algorithms such as navigable small world graphs (NSW) and hierarchical NSW (HNSW) when dealing with high dimensional datasets, as these methods can provide faster and more accurate results than traditional exact nearest neighbor search techniques. (Cayton 2008)

  • Consider utilizing scalable nearest neighbor algorithms for high dimensional data analysis, particularly when dealing with large datasets, as these algorithms offer improved efficiency and accuracy over traditional methods. (NA?)

  • Carefully consider the choice of k value when applying kNN classification, as different test data points might require different numbers of nearest neighbors for accurate predictions. (NA?)

  • Focus on developing efficient and accurate approximate nearest neighbor search algorithms for billion-scale datasets, while considering hardware costs and limitations. (NA?)

  • Carefully select the appropriate distance measure for your KNN classifier, as the choice of distance measure significantly influences the classifiers performance across various datasets and noise levels.’ (NA?)

  • Carefully select and evaluate similarity and distance metrics for k-nearest neighbor classifiers, considering factors like data type, computational efficiency, and domain knowledge, to improve classification accuracy and efficiency. (NA?)

Principal Component Analysis (Pca)

  • Utilise an acceleration scheme for memory-limited streaming PCA that doesn’t need any pre-defined parameters or pre-processing steps. (Alakkari and Dingliana 2018)

  • Use Principal Component Analysis (PCA) to identify the most meaningful basis to re-express your dataset, aiming to filter out noise and reveal hidden structures. (Shlens 2014)

  • Consider utilizing nonstandard inner products or metrics in analyzing high-dimensional data, as this is a straightforward way to integrate complex external information like graphical data into the analysis. (Purdom 2011)

  • Utilize the proposed algorithm for performing Principal Component Analysis (PCA) on tree-structured data, which offers a computationally efficient and accurate way to analyze complex datasets. (Aydın et al. 2009)

  • Utilise the novel kernel function f(A,B) derived from the concept of principal angles between two linear subspaces, which enables accurate comparison of sets of vectors while maintaining computational efficiency through the use of inner-products between pairs of column vectors. (NA?)

  • Use matrix perturbation theory and concentration of measure bounds on the norm of noisy Wishart matrices to understand the differences between sample principal component analysis (PCA) and population PCA, particularly in cases where the number of dimensions (p) and the number of samples (n) are large. (NA?)

  • Utilise the truncated power method’, a novel approach to solving the sparse eigenvalue problem, which involves applying the standard power method to a sparse eigenvector whilst ensuring sparsity throughout the process.’ (NA?)

  • Carefully select appropriate statistical techniques for metabolomics data analysis, taking into account potential issues such as high data dimensionality, over-fitting, and limitations of individual methods, and consider combining different analytical technologies and statistical tools for a comprehensive interpretation of results. (NA?)

  • Utilize matrix sketching techniques like Frequent Directions and random projections to develop efficient and practical algorithms for anomaly detection in high-dimensional data, achieving space linear or sublinear in the dimension. (NA?)

Unsupervised Learning Algorithms

Hierarchical Clustering

  • Consider using partitional clustering algorithms for clustering large document datasets, as they offer lower computational requirements and often perform better than agglomerative algorithms. (Ying Zhao and Karypis 2002)

  • Focus on developing incremental algorithms for efficiently building concept lattices, which can effectively organize and represent complex data structures for various applications, such as information retrieval and browsing. (NA?)

  • Carefully consider the choice of constraints when applying agglomerative hierarchical clustering methods, as some constraint combinations can lead to NP-complete feasibility problems and dead-end solutions, while others can improve cluster purity and average distortion. (NA?)

  • Consider using hierarchical clustering methods, specifically single linkage hierarchical clustering, due to its unique characteristics, stability, and convergence properties demonstrated through the use of the Gromov-Hausdorff distance. (NA?)

  • Focus on developing a representative trace sampling technique that selects a diverse range of execution traces, rather than relying solely on traditional statistical techniques such as stratified sampling or importance sampling. (NA?)

  • Consider using hierarchical clustering methods, specifically single linkage hierarchical clustering, due to its unique characteristics, stability, and convergence properties, especially when dealing with complex datasets with multiscale structures. (NA?)

K-Means Clustering

  • Avoid relying solely on the traditional k-means algorithm for solving the Minimum Sum-of-Squares Clustering (MSSC) problem, as it is proven to be NP-hard for k = 2 and general dimensions through a valid reduction from the densest cut problem. (NA?)

  • Consider using an adjusted version of iK-Means for optimal performance in determining the right number of clusters, cluster recovery, and centroid recovery in K-Means clustering. (NA?)

Latent Dirichlet Allocation (Lda)

  • Carefully consider and account for finite-sample bias when analyzing high-dimensional choices, as failing to do so can lead to misleading conclusions about group differences. (Gentzkow, Shapiro, and Taddy 2019)

  • Carefully consider the interplay of multiple factors, such as the number of documents, document length, number of topics, and Dirichlet hyperparameters, when applying latent Dirichlet allocation (LDA) models to your datasets, as these factors significantly impact the models performance and accuracy.’ (Animashree Anandkumar, Hsu, and Kakade 2012)

  • Consider using generative statistical topic models for multi-label document classification, especially when dealing with large numbers of relatively rare labels and skewed label frequencies, as these models can achieve competitive performance compared to discriminative methods while providing explicit assignments of individual words to specific labels and jointly modeling all labels within a corpus. (Rubin et al. 2011)

  • Consider decomposing traditional random walk into multiple random walks specific to various topics when conducting keyphrase extraction, as this allows for improved accuracy and comprehensiveness in capturing the main ideas of a text. (NA?)

  • Consider using the RPC-based change point method to determine the optimal number of topics in topic modeling, as it demonstrates improved stability and effectiveness compared to traditional perplexity-based methods. (NA?)

Non-Negative Matrix Factorization (Nmf)

  • Utilise a two-layer strategy for applying topic modeling in a non-negative matrix factorisation framework to a timestamped corpus of political speeches. (Greene and Cross 2016)

  • Consider using Poincare embeddings for learning hierarchical representations of symbolic data, as they allow for simultaneous capture of hierarchy and similarity, leading to improved representation capacity and generalization ability compared to traditional Euclidean embeddings. (Bojanowski et al. 2016)

  • Use Bayesian non-negative matrix factorization (NMF) with a Gibbs sampler for improved interpretability and uncertainty estimation, along with model order selection via marginal likelihood estimation and computation of the maximum a posteriori (MAP) estimate using an iterated conditional modes algorithm. (“Independent Component Analysis and Signal Separation” 2009)

  • Consider using sparse non-negative matrix factorization (SNMF) as a computationally efficient approach to separate speech sources in a single-channel recording, while exploring the impact of varying degrees of sparseness and the number of dictionary elements. (NA?)

Independent Component Analysis (Ica)

  • Use contrastive learning techniques to implicitly invert the underlying generative model of your observed data, leading to improved generalizability and effectiveness in various downstream tasks. (Zimmermann et al. 2021)

Anomaly Detection

  • Carefully consider the availability of data labels and choose the appropriate anomaly detection setting accordingly, whether it be unsupervised, semi-supervised, or supervised, to effectively capture the nuances of the anomalies being studied. (Hojjati, Ho, and Armanfard 2024)

  • Use the Elliptic++ dataset, which combines the Elliptic dataset and the Bitcoin addresses dataset, to analyze and detect fraudulent activities in cryptocurrency transactions. (Elmougy and Liu 2023)

  • Differentiate structural patterns for anomalies and normals in order to alleviate structural distribution shifts (SDS) in graph anomaly detection (GAD). (Yuan Gao et al. 2023)

  • Consider using Midas-F, a modified version of the Midas algorithm, to improve the accuracy of your anomaly detection models in edge streams by reducing the “poisoning” effect caused by incorporating anomalies into the algorithms internal states.’ (Siddharth Bhatia et al. 2022)

  • Carefully select datasets, model parameters, and evaluation measures to minimize biases and accurately assess the performance of time-series anomaly detection methods. (Paparrizos et al. 2022)

  • Consider employing model-independent tests alongside model-dependent ones when searching for new physics signals, as these tests offer increased flexibility and robustness against mis-specifications in the signal model. (Chakravarti et al. 2021)

  • Carefully choose appropriate methods for anomaly detection and root cause analysis in multi-service applications, taking into account factors such as the level of detail required, the complexity of the application, and the available data sources. (Soldani and Brogi 2021)

  • Consider using deep reconstruction techniques for unsupervised video anomaly detection (UVAD), as these methods offer a normality advantage’, allowing for easier identification of anomalies within unlabelled videos.’ (G. Yu et al. 2021)

  • Explicitly learn the low-dimensional intermetric and temporal representations with appropriate structural designs to effectively capture the normal patterns of multi-time series (MTS) data for improved anomaly detection and interpretation. (Zhihan Li et al. 2021)

  • Utilise the SAND methodology for subsequence anomaly detection in data streams, which involves updating a weighted set of subsequences over time using k-Shape clustering algorithm, merging similar clusters, and calculating anomaly scores based on the current batch. (Boniol et al. 2021)

  • Use a behaviour-driven taxonomy when defining time series outliers, which allows for clearer context definitions and more effective synthesis of different types of outliers for benchmarking purposes. (Angryk et al. 2020)

  • Interpret the density level detection problem as a binary classification problem, allowing them to utilise the corresponding empirical classification risk as an empirical performance measure for anomaly detection. (DeSantis et al. 2020)

  • Consider using a data-dependent point kernel when implementing kernel mean embeddings for anomaly detection, as it addresses the issues of intractable dimensionality and data independence present in traditional approaches. (Ting et al. 2020)

  • Utilize a loss function framework for calibrated anomaly detection, which enables them to estimate the density for anomalous instances while ignoring the density for non-anomalous instances, providing implicit quantile control through a connection to the generalized pinball loss, and offering efficient optimization with kernelized scores for a specific family of losses. (Tuson et al. 2020)

  • Consider using the Series2Graph algorithm for unsupervised subsequence anomaly detection, which employs a graph representation of low-dimensionality embeddings of subsequences to distinguish between normal and anomalous patterns without requiring labeled instances or clean data. (Boniol and Palpanas 2020)

  • Carefully differentiate between the terms “rare event,” “anomaly,” “novelty,” and “outlier” when conducting studies involving abnormal observations, as each term refers to a unique learning scenario within the broader context of supervised classification. (Carreño, Inza, and Lozano 2019)

  • Use a graph-based sampling and consensus (GraphSAC) approach to effectively detect anomalous nodes in large-scale graphs, rather than relying solely on connectivity and attributes of all nodes, which could be compromised by adversaries. (Ioannidis, Berberidis, and Giannakis 2019)

  • Consider the impact of log instability on the effectiveness of log-based anomaly detection approaches, particularly in terms of handling evolving logging statements and processing noise. (Xu Zhang et al. 2019)

  • Consider using the Spectral Residual (SR) algorithm for unsupervised time-series anomaly detection due to its simplicity, efficiency, and effectiveness, particularly when dealing with large amounts of data without labeled examples. (H. Ren et al. 2019)

  • Consider employing a unified data-driven deep-learning framework like LogAnomaly for accurate and efficient anomaly detection in unstructured log streams, which addresses the limitations of traditional methods by incorporating semantic information through template2Vec and simultaneous detection of sequential and quantitative anomalies. (Weibin Meng et al. 2019)

  • Use a set-based update approach instead of a point-based update approach for distance-based outlier detection in data streams, as it allows for efficient grouping of close data points and extraction of net changes between expired and new data points, ultimately reducing unnecessary update operations and improving computational efficiency. (S. Yoon, Lee, and Lee 2019)

  • Focus on developing a unified framework for detecting out-of-distribution samples and adversarial attacks, leveraging the power of machine learning algorithms like deep neural networks and statistical tools like the Mahalanobis distance. (Kimin Lee et al. 2018)

  • Carefully balance the tradeoff between reducing false positives and false negatives in anomaly detection, particularly when dealing with multivariate time series data, and consider using advanced techniques like LSTM recurrent neural networks combined with nonparametric, dynamic, and unsupervised thresholding approaches to achieve high prediction performance while maintaining interpretability. (Hundman et al. 2018)

  • Consider integrating representation learning and outlier detection processes within a unified framework, such as RAMODO, to optimize the performance of distance-based outlier detection algorithms in ultrahigh-dimensional datasets. (G. Pang et al. 2018)

  • Leverage the power of online convex optimization to create efficient and effective algorithms for feedback-guided anomaly detection, resulting in significant improvements in accuracy and speed. (Siddiqui et al. 2018)

  • Consider a comprehensive acceleration framework called SUOD when dealing with large-scale heterogeneous outlier detection, as it addresses issues related to data reduction, model approximation, and task load balance in a distributed environment. (Aggarwal and Sathe 2017)

  • Focus on developing a comprehensive understanding of the underlying mechanisms driving the phenomenon of interest, rather than relying solely on traditional statistical techniques. (Shenghua Liu, Hooi, and Faloutsos 2017)

  • Develop a comprehensive understanding of the outlier detection process by breaking down the overall problem into multiple regional tasks of explaining individual outliers, allowing for more accurate and efficient interpretation of outlier detection results. (N. Liu, Shin, and Hu 2017)

  • Carefully consider the type of anomaly they want to detect (point vs. group), the format of the input data (activity vs. graph), and the role of temporality in the social network when developing and evaluating social media anomaly detection methods. (R. Yu et al. 2016)

  • Utilise a network diffusion based framework to identify significant causal anomalies and rank them, rather than just focusing on the percentage of vanishing correlations. This approach allows for more accurate modelling of fault propagation across the entire invariant network, and performs joint inference on both the structural and time-evolving broken invariance patterns, leading to improved identification of high-confidence anomalies and compensation for unstructured measurement noise in the system. (W. Cheng et al. 2016)

  • Consider the effects of high dimensionality on unsupervised outlier-detection methods and hubness, explore the relationship between hubness and data sparsity, and develop methods like AntiHub and AntiHub² to improve discrimination of scores in outlier detection. (Radovanovic, Nanopoulos, and Ivanovic 2015)

  • Consider combining reverse distillation with multi-task learning to enhance feature compactness and anomalous signal suppression, leading to improved performance in anomaly detection tasks. (G. Hinton, Vinyals, and Dean 2015)

  • Carefully consider the unique characteristics of your temporal dataset when selecting appropriate outlier detection techniques, taking into account factors such as data type, supervision, and the specific challenges posed by temporal data. (M. Gupta et al. 2014)

  • Use a diverse set of benchmark datasets, carefully controlled along four problem dimensions - point difficulty, semantic variation, relative frequency, and feature relevance/irrelevance - to accurately evaluate and compare the performance of anomaly detection algorithms. (Alzghoul and Löfstrand 2011)

  • Consider using semi-supervised learning techniques, specifically the ADS framework, for rapid deployment of anomaly detection models across large numbers of emerging KPI streams, without requiring manual algorithm selection, parameter tuning, or new anomaly labeling for each new stream. (Mahimkar et al. 2011)

  • Utilise the TOD framework, which employs a novel programming model that breaks down a variety of OD applications into a small collection of basic tensor operators and functional operators, thereby reducing implementation and optimization effort and allowing for easier incorporation of new OD algorithms. (Alshawabkeh, Jang, and Kaeli 2010)

  • Carefully consider the nature of the input data, the type of anomaly, and the application domain when selecting and developing appropriate anomaly detection techniques. (Chandola, Banerjee, and Kumar 2009)

  • Utilize active learning to detect both errors and events in time series data, employing a non-parametric concept of neighborhood and probabilistic classification to achieve accurate labelling with minimal user interaction. (R. P. Adams and MacKay 2007)

  • Focus on developing machine learning techniques to effectively differentiate normal and abnormal patterns in Unix process execution traces for improved intrusion detection. (W. Lee, Stolfo, and Chan 1997)

  • Consider applying advanced statistical signal processing techniques, specifically those focused on abrupt change detection, to the problem of network anomaly detection in order to gain deeper insights and potentially improve the reliability of IP networks. (NA?)

  • Carefully select an appropriate outlier detection methodology based on factors such as data distribution, dimensionality, and desired accuracy, considering options ranging from proximity-based techniques like k-nearest neighbor to advanced machine learning algorithms. (NA?)

  • Consider using machine learning techniques to analyze console logs, specifically focusing on creating features that capture correlations between different types of log messages, in order to improve the accuracy of problem detection in large-scale data center services. (NA?)

  • Consider using one-class SVMs for unsupervised anomaly detection, especially the eta one-class SVM, as it consistently outperforms other algorithms in terms of accuracy and sparsity. (NA?)

  • Carefully choose among various novelty detection methods, including probabilistic, distance-based, reconstruction-based, domain-based, and information-theoretic techniques, considering factors like accuracy, computational cost, and applicability to specific domains. (NA?)

  • Develop a comprehensive benchmarking system like NAB to evaluate real-time anomaly detection algorithms on streaming data, taking into account factors such as early detection, false alarm minimization, and adaptation to changing statistical patterns. (NA?)

  • Consider employing deep learning techniques for anomaly detection, as they allow for end-to-end optimization of the anomaly detection pipeline and the ability to learn representations specifically tailored for anomaly detection, thereby improving the utilization of limited labeled data and increasing overall detection accuracy. (NA?)

  • Develop a comprehensive taxonomy to classify existing anomaly detection algorithms for building energy consumption based on different modules and parameters, such as machine learning algorithms, feature extraction approaches, anomaly detection levels, computing platforms, and application scenarios. (NA?)

  • Develop an efficient, unsupervised method for detecting anomalies in edge streams using a novel frequency-factorization technique that takes advantage of both temporal and structural information in a streaming manner, allowing for constant memory usage. (NA?)

Recommendation Systems

  • Use causal inference techniques to overcome the limitations of traditional correlation-based recommender systems, particularly in terms of data bias, missing data, and beyond-accuracy objectives. (C. Gao et al. 2024)

  • Consider integrating Large Language Models (LLMs) into your recommender systems to improve interactivity, explainability, and cross-domain recommendations, while also addressing the cold-start problem. (Yunfan Gao et al. 2023)

  • Systematically investigate how to extract and transfer knowledge from pre-trained models learned by different PLM-related training paradigms to improve recommendation performance from various perspectives, such as generality, sparsity, efficiency and effectiveness. (Peng Liu, Zhang, and Gulla 2023)

  • Focus on developing a generative recommendation system that uses artificial intelligence to create personalized content tailored to individual users needs, while incorporating user instructions to guide the content generation process.’ (Wenjie Wang et al. 2023)

  • Carefully consider the choice of recommendation architecture and training approach when comparing MoRec and IDRec, as the effectiveness of MoRec depends heavily on these factors. (Zheng Yuan, Yuan, Song, et al. 2023)

  • Carefully consider and account for confounding variables in your analysis, particularly those that might introduce spurious correlations or lead to biased recommendations. (Xiangnan He et al. 2022)

  • Use a two-step item representation scheme (“text -> code -> representation”) to improve the accuracy of your sequential recommenders, rather than directly mapping text encodings into item representations. This allows for greater flexibility in fitting new recommendation scenarios while reducing the influence of text semantics on the recommendation model. (Y. Hou et al. 2022)

  • Employ multi-objective hyper-parameter optimization to avoid negative side effects caused by single-objective optimization in behavioral song embeddings. (Quadrana, Larreche-Mouly, and Mauch 2022)

  • Consider incorporating personalized prompt-based recommendation (PPR) frameworks when working with pre-trained recommendation models, as they allow for more efficient extraction of relevant knowledge from these models, particularly in cold-start scenarios. (Yiqing Wu et al. 2022)

  • Consider using the CSP dataset, which includes user profiles and item ratings, to develop and test algorithms for solving the cold start problem in recommender systems. (Herce-Zelaya et al. 2022)

  • Carefully select your data sources and preprocess them appropriately, considering factors such as missing values, outliers, and variable transformations, to ensure accurate and reliable results when applying statistical methods. (Kreutz and Schenkel 2022)

  • Consider utilizing Monte Carlo Tree Search (MCTS) for dynamic selection of items to present to new users in a recommender system, as it demonstrates faster and more accurate identification of user preferences compared to traditional decision-trees or state-of-the-art bandit-based approaches, while maintaining computational efficiency. (Rajapakse and Leith 2022)

  • Consider incorporating future-aware diverse trends (FAT) framework into your sequential recommendation systems to improve the accuracy and diversity of recommendations by capturing the evolving preferences of users over time. (Yujie Lu et al. 2021)

  • Account for the exposure model bias in classical matrix factorization by using the deconfounded recommender, which utilizes an exposure model to estimate a substitute for unobserved confounders and subsequently fits a ratings model that accounts for those substitutes. (“Fourteenth ACM Conference on Recommender Systems” 2020)

  • Consider utilising a controllable multi-interest framework when developing sequential recommendation systems, as it allows for better reflection of a users multiple interests during a given period, leading to improved recommendation accuracy and diversity.’ (Cen et al. 2020)

  • Utilise the Linear Modular Dispersion Bandit (LMDB) framework for diversified interactive recommendation, which combines modular functions for relevance properties and dispersion functions for diversity properties, allowing for a balance between recommendation accuracy and diversity. (Q. Ding et al. 2020)

  • Consider using Deep Retrieval (DR) for large-scale recommendation tasks, as it allows for efficient and accurate retrieval of top relevant candidates without relying on the Euclidean space assumption inherent in traditional Approximate Nearest Neighbor (ANN) algorithms. (W. Gao et al. 2020)

  • Incorporate self-supervised learning (SSL) into your recommendation systems to enhance item representation learning, particularly for long-tail distributions and sparse data, by employing a novel data augmentation method that exploits feature correlations and is tailored for heterogeneous categorical features commonly found in recommender models. (T. Yao et al. 2020)

  • Consider utilising a controllable multi-interest framework when developing sequential recommendation systems, as it allows for better reflection of a users multiple interests during a given period, leading to improved recommendation accuracy and diversity.’ (Cen et al. 2020)

  • Employ multi-objective hyper-parameter optimization to avoid negative side effects caused by single-objective optimization in behavioral song embeddings. (Pal et al. 2020)

  • Carefully examine and reconsider the underlying assumptions about product relationships in co-purchase and co-view data, and propose a new approach to collect labels as distant supervision for complementary product recommendation (CPR). (Junheng Hao et al. 2020)

  • Develop a query understanding module to expose under-served content in search results, leveraging various features like standalone, reference-dependent, and interaction-based features to accurately identify non-focused queries suitable for surfacing under-served content. (Tomasi et al. 2020)

  • Consider using the Zero-Shot Heterogeneous Transfer Learning framework to improve search retrieval performance by leveraging (item, item) correlations within a recommender system. (T. Wu et al. 2020)

  • Focus on developing unbiased and fair dynamic learning-to-rank algorithms to address issues of bias and unfairness in various domains, such as news aggregation platforms, job applicant ranking systems, and online marketplaces. (Morik et al. 2020)

  • Consider using a multi-interest network with dynamic routing (MIND) when attempting to model diverse user interests in the matching stage of industrial recommender systems, as demonstrated by its superior performance compared to state-of-the-art methods in extensive experiments on public benchmarks and a large-scale industrial dataset from Tmall. (Chao Li et al. 2019)

  • Utilise a sequential deep matching (SDM) model to improve the accuracy of predictions in large-scale recommender systems by modelling both short-term sessions and long-term behaviours. (F. Lv et al. 2019)

  • Carefully consider the appropriate balance between accuracy and diversity when developing content-based recommendation algorithms, taking into account factors such as interaction frequency and recency, and potentially leveraging models like Base-Level Learning (BLL) to optimize your approach. (Reiter-Haas et al. 2019)

  • Consider using a combination of Graph Convolutional Networks (GCNs) and knowledge distillation techniques to improve the efficiency and effectiveness of top-K item recommendations based on implicit feedback. (Haoyu Wang, Lian, and Ge 2019)

  • Consider integrating future data into model training for session-based recommendation systems, despite the challenge of avoiding data leakage, as it provides valuable signals about user preferences and can enhance recommendation quality. (F. Yuan et al. 2019)

  • Utilise knowledge-aware graph neural networks with label smoothness regularisation’ (KGNN-LS) to enhance the accuracy of your predictions in recommender systems. (Hongwei Wang, Zhang, et al. 2019)

  • Consider using collaborative filtering techniques for time-constrained model selection and hyperparameter tuning in automated machine learning (AutoML) projects, as demonstrated by the successful implementation of the Oboe algorithm. (Chengrun Yang et al. 2019)

  • Utilise a sequential deep matching (SDM) model to improve the accuracy of predictions in large-scale recommender systems by modelling both short-term and long-term user behaviour patterns. (F. Lv et al. 2019)

  • Consider combining generalized linear mixed (GLMix) models with gradient boosted decision tree (GBDT) models to achieve entity-personalized talent search recommendations through nonlinear tree interaction features. (Ozcaglar et al. 2019)

  • Be aware of the potential for unbalanced recommendations when optimizing for accuracy in recommendation systems, and consider incorporating calibration techniques to mitigate this issue. (“Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence” 2018)

  • Consider employing a combination of topic modelling and multi-armed bandits techniques to optimize candidate quality in talent search recommendation systems, balancing exploration and exploitation to effectively adapt to individual user preferences. (Geyik, Dialani, et al. 2018)

  • Consider the unique challenges posed by talent search and recommendation systems, including handling complex queries, measuring mutual interest between recruiters and candidates, and supporting search based on ideal candidates, when developing and deploying such systems. (Geyik, Guo, et al. 2018)

  • Carefully consider the role of sequential patterns and order constraints in developing effective sequence-aware recommendation systems, taking into account the unique challenges posed by different application scenarios. (Quadrana, Cremonesi, and Jannach 2018)

  • Consider incorporating higher-order item relations into your item-based collaborative filtering models to enhance the accuracy and interpretability of user preference modeling. (F. Xue et al. 2018)

  • Develop a two-stage recommender system consisting of a matching phase followed by a ranking phase, with the matching phase focusing on computing pairwise similarities between items based on users behaviors, and the ranking phase leveraging deep neural networks to rank candidate items according to individual user preferences.’ (Jizhe Wang et al. 2018)

  • Consider utilizing a combination of graph pruning techniques, multi-query pin analysis, and early stopping strategies to improve the efficiency and accuracy of your recommendation algorithms. (Eksombatchai et al. 2017)

  • Carefully analyze and reconsider the assumptions underlying traditional co-purchase and co-view data, and instead adopt a more nuanced approach to collecting and interpreting these data points, enabling more accurate and comprehensive understanding of complementary product relationships. (Hamilton, Ying, and Leskovec 2017b)

  • Consider utilizing a mixture model approach for predicting push message open rates in e-commerce settings, where the model learns latent prediction contexts based on user and item profiles, and optimizes its parameters using an expectation-maximization algorithm. (H. Zhao et al. 2017)

  • Consider using deep neural networks to model collaborative filtering effects in recommender systems, particularly for handling noisy implicit feedback signals. (Xiangnan He et al. 2017)

  • Consider adopting a translation-based model for sequential recommendation tasks, as it effectively captures personalized sequential behavior while scaling efficiently to large, real-world datasets. (R. He, Kang, and McAuley 2017)

  • Utilize deep learning techniques when dealing with large-scale recommendation systems like YouTube, as it offers significant performance improvements compared to traditional matrix factorization methods. (Covington, Adams, and Sargin 2016)

  • Utilize the Fully Coupled Interaction Tensor Factorization (FCTF) model for estimating Click Through Rates (CTR) in Real Time Bidding (RTB) display advertising. This model effectively handles the complex interactions among user, publisher, and advertiser, while maintaining linear runtime complexity for both learning and prediction. (Shan et al. 2016)

  • Consider extending existing item-based recommendation models to incorporate both global and local item-item models, which can capture differences in user preferences and potentially improve top-N recommendation performance. (Christakopoulou and Karypis 2016)

  • Develop a flexible framework for optimizing whole-page presentation of search results, taking into consideration various aspects like item positions, image sizes, text fonts, and other styles, while being mindful of business and design constraints. (Yue Wang et al. 2016)

  • Utilize A/B testing alongside offline experimentation using historical member engagement data to improve recommendation algorithms, while considering potential biases and limitations inherent in these methods. (Gomez-Uribe and Hunt 2015)

  • Conduct a systematic review of the literature to identify the most commonly used machine learning algorithms in recommender systems and explore opportunities for improvement through software engineering practices. (Portugal, Alencar, and Cowan 2015)

  • Leverage the power of neural language models to create low-dimensional, distributed embeddings of products from historical logs, enabling accurate product recommendations through simple nearest neighbor search. (Grbovic et al. 2015)

  • Use a Bayesian approach called “HypTrails”, which combines Markov chain modelling and Bayesian inference, to effectively compare multiple hypotheses about human trails on the web. (P. Singer et al. 2015)

  • Employ a combination of data hierarchy and temporal smoothing techniques to improve the accuracy of click-through rate (CTR) estimation for rare events in online advertising. (G. Sun, Bin, and Zhou 2015)

  • Develop a greedy interaction feature selection algorithm based on gradient boosting to identify and prioritize relevant interaction features in context-aware recommendation systems, thereby improving the overall performance of the recommendation model. (C. Cheng et al. 2014)

  • Consider using dwell time as a proxy for user satisfaction in personalized content recommendation systems, as it provides a more comprehensive measure of user engagement than traditional click-through rates. (X. Yi et al. 2014)

  • Carefully choose the appropriate loss function when training large-scale factorized recommendation models, considering whether to prioritize higher or lower ranked items depending on the desired outcome, such as improved precision and recall metrics or better mean or maximum rank metrics. (Weston, Yee, and Weiss 2013)

  • Consider implementing a focused matrix factorization (FMF) approach when dealing with sparse user-product interaction matrices in order to improve audience selection accuracy in display advertising. (Kanagal et al. 2013)

  • Focus on developing tag-aware recommender systems using network-based, tensor-based, and topic-based methods to effectively leverage the wealth of information contained in social tagging systems. (Z.-K. Zhang, Zhou, and Zhang 2011)

  • Carefully analyze the distribution of ratings and user behavior patterns in large-scale music datasets, taking into consideration factors such as user activity levels, item popularity, and temporal effects, in order to develop effective predictive models for music recommendation systems. (Cremonesi, Koren, and Turrin 2010)

  • Carefully choose appropriate collaborative filtering techniques depending on specific challenges faced, such as data sparsity, scalability, synonymy, gray sheep, shilling attacks, personal privacy, explainability, and noise, and utilize various strategies including dimensionality reduction, hybrid algorithms, and model-based approaches to optimize recommendation performance. (X. Su and Khoshgoftaar 2009)

  • Consider using a user profile merging technique based on total distance minimization to effectively combine individual user preferences and generate a common program recommendation list for multiple viewers. (Z. Yu et al. 2006)

  • Consider combining traditional feature reduction techniques with collaborative filtering methods to enhance the quality of feature spaces in text classification tasks, leading to improved classification accuracy. (Yang Song et al. 2006)

  • Focus on developing advanced user models that take into account a wide range of factors, including cognitive abilities, individual differences, behavior patterns, subject domains, work tasks, work environments, and temporal dynamics, while also considering the community aspect of human behavior and preferences. (Smeaton and Callan 2005)

  • Utilise item-to-item collaborative filtering when dealing with large datasets, as it allows for efficient and accurate recommendations without compromising on quality. (Linden, Smith, and York 2003)

  • Consider implementing content-based book recommending systems that utilize information extraction and machine-learning algorithms for text categorization, as opposed to solely relying on collaborative filtering methods, in order to accurately recommend previously unrated items to users with unique interests while providing explanations for those recommendations. (NA?)

  • Incorporate dynamic features, such as instantaneous click-through rates, into your content feature sets and continuously update these dynamic features in real-time using user interactions to effectively address the cold-start problem in personalized recommendation systems. (NA?)

  • Use a rating-matrix generative model (RMGM) for effective cross-domain collaborative filtering, which involves combining a user-item joint mixture model and a cluster-level rating model to fill missing ratings for both existing and new users. (NA?)

  • Utilize attribute-to-feature mappings within matrix factorization models to improve the predictive accuracy of cold-start recommendations, while maintaining computational efficiency. (NA?)

  • Carefully construct your training and testing datasets using multiple random sampling methods to ensure accurate evaluation of model performance and avoid bias. (NA?)

  • Consider utilizing Factorization Machines (FMs) for context-aware rating predictions due to your ability to compute the model equation in linear time, making them faster and more efficient than existing methods like Multiverse Recommendation. (NA?)

  • Expand your evaluation criteria beyond simple prediction accuracy to encompass various aspects of user experience, such as user engagement, trust, and satisfaction, as well as the potential tradeoffs between individual and collective benefits within a recommender system. (NA?)

  • Consider integrating exogenous knowledge from the Linked Open Data (LOD) cloud into your graph-based recommendation frameworks, as doing so can significantly enhance the performance of these systems. (NA?)

  • Focus on developing a hybrid approach that combines the strengths of both campaign-agnostic and campaign-aware expansion techniques to maximize the efficiency and effectiveness of audience expansion strategies in online social network advertising. (NA?)

  • Conduct a systematic review of the literature to identify trends in the use or research of machine learning algorithms in recommender systems, identify open questions in the use or research of machine learning algorithms, and assess new researchers to position new research activity in this domain appropriately. (NA?)

  • Ensure your studies are reproducible by providing open-access source code and data, and conduct rigorous testing against multiple baselines to demonstrate the effectiveness of your proposed deep learning algorithms for top-n recommendations. (NA?)

  • Utilize a pairwise recommendation fairness metric to evaluate the fairness of a recommender system, which involves running randomized experiments to obtain unbiased estimates of user preferences and subsequently applying a novel regularization technique to enhance the ranking fairness of the pointwise recommender system. (NA?)

  • Consider leveraging large language models (LLMs) as a general-purpose recommendation tool for handling multiple recommendation tasks, including rating prediction, sequential recommendation, direct recommendation, explanation generation, and review summarization, potentially improving recommendation performance by tapping into the extensive linguistic and world knowledge acquired from large-scale corpora. (NA?)

  • Adopt a prompt learning paradigm with a two-phase training strategy for non-overlapping many-to-one cross-domain recommendation (NMCR) tasks, incorporating domain-agnostic and domain-specific prompts to effectively transfer domain knowledge from multiple source domains to the target domain while preserving its unique characteristics. (NA?)

  • Consider using large language models (LLMs) for graph augmentation in recommendation systems, specifically focusing on augmenting user-item interaction edges, item node attributes, and user node profiles, while also developing a denoised data robustification mechanism to maintain the quality of the augmented data. (NA?)

Reinforcement Learning

Temporal Difference Learning

  • Consider using temporal-difference (TD) methods instead of traditional supervised-learning methods for prediction tasks, especially in situations involving multi-step prediction problems, as TD methods offer significant computational efficiency benefits and often lead to faster convergence and improved accuracy. (NA?)

Q-Learning

  • Consider using dynamic programming and Bayesian methods to optimize reinforcement learning tasks, allowing for smoother movement from exploratory to exploitative behaviors and improved overall performance. (NA?)

Monte Carlo Tree Search (Mcts)

  • Utilise the Factored Bayes-Adaptive POMDP model for reinforcement learning tasks in partially observable systems, as it effectively balances exploration and exploitation, learns the dynamics of the system, and scales well even in complex scenarios. (Katt, Oliehoek, and Amato 2018)

  • Utilise Monte Carlo Tree Search (MCTS) to learn about vine copula structures. (Kraus and Czado 2017)

  • Consider offline learning of partial policies to reduce action branching in large MDPs, thereby improving the efficiency of tree search methods like Monte Carlo tree search without compromising decision-making quality. (V. Mnih et al. 2013)

  • Utilise Monte Carlo Tree Search (MCTS) in conjunction with deep neural networks to effectively and efficiently discover retrosynthetic routes for chemical syntheses. (NA?)

Deep Reinforcement Learning

  • Develop a deep reinforcement learning-enabled algorithm for optimal Artificial Intelligence-Generated Content (AIGC) Service Provider (ASP) selection in order to maximize the quality of generated content in wireless edge networks. (H. Du et al. 2023)

  • View prompt optimization as a strategic planning problem and use a principled planning algorithm, such as Monte Carlo tree search, to navigate the vast space of expert-level prompts. (Xinyuan Wang et al. 2023)

  • Consider implementing a three-step prompting strategy for GPT-3 to effectively make next-item recommendations in a zero-shot setting, which involves capturing user preferences, selecting representative movies, and generating a ranked list of 10 movies, ultimately leading to enhanced recommendation accuracy. (Lei Wang and Lim 2023)

  • Consider combining foundation models and sequential decision making techniques to enhance your ability to handle complex tasks involving long-term reasoning, control, search, and planning, thereby improving overall performance and generalizability. (S. Yang et al. 2023)

  • Consider converting multi-task learning problems into learning basic skills and planning over those skills, particularly in open-world environments where traditional reinforcement learning is highly inefficient. (Haoqi Yuan et al. 2023)

  • Carefully consider the complexities involved in applying reinforcement learning techniques to cellular network configuration problems, including the large number of configuration parameters, difficulty specifying contexts, and risk of performance degradation. (Changhan Ge et al. 2023)

  • Explore the relationship between parameters and performance in Hyperledger Fabric, a permissioned blockchain, and develop an auto-tuning system like Athena that employs a Permissioned Blockchain Multi-Agent Deep Deterministic Policy Gradient (PB-MADDPG) to achieve heterogeneous parameter-tuning optimization across different types of nodes in Fabric. (Mingxuan Li et al. 2023)

  • Consider combining multiple pre-trained models, specifically those related to language, vision, and action, to develop effective systems for robotic navigation that can interpret and respond to natural language instructions. (Wenlong Huang et al. 2022)

  • Consider using a combination of explicit motion parameterization, simultaneous learning of motion parameterization and motor skills, continuous-time reinforcement learning, and automatic curriculum generation to improve the performance and visual quality of learned motor skills. (Seyoung Lee et al. 2021)

  • Use the Go-Explore algorithm to address the challenges of detachment and derailment in reinforcement learning, enabling thorough exploration of environments and leading to significant performance improvements. (Ecoffet et al. 2021)

  • Utilize a deep reinforcement learning (RL) model called Partner to enhance empathy in online mental health support conversations by making sentence-level edits to posts, thereby increasing the expressed level of empathy while preserving conversation quality. (Sharma et al. 2021)

  • Leverage the structure of Attribute Dynamic Graphs (ADGs) and historical experience (traversed temporal paths) to improve the accuracy of state estimation and increase the frequency of positive rewards in the context of multi-constrained temporal path discovery. (Pengfei Ding et al. 2019)

  • Consider employing deep reinforcement learning (DRL) for power system emergency control due to its ability to enable automatic high-dimensional feature extraction and end-to-end learning through stochastic gradient descent, improving scalability and suitability for solving large-scale control problems. (Qiuhua Huang et al. 2019)

  • Consider utilizing multiple policy value neural networks (PV-NNs) of varying sizes within a Monte Carlo tree search (MCTS) framework to optimize game playing performance, as demonstrated by the superior performance of the multiple policy value MCTS (MPV-MCTS) approach over traditional single PV-NN with policy value MCTS (PV-MCTS) in the game NoGo. (L.-C. Lan et al. 2019)

  • Consider applying deep reinforcement learning (DRL) to dynamic pricing problems in e-commerce, particularly by extending the discrete set problem to the continuous price set, defining a new reward function named difference of revenue conversion rates (DRCR), and tackling the cold-start problem of MDP through pre-training and evaluation using historical sales data. (Jiaxi Liu et al. 2019)

  • Consider the unique challenges posed by multi-agent reinforcement learning (MARL) compared to traditional single-agent reinforcement learning (RL), such as the need to account for multiple dimensions of learning goals, non-stationary environments, combinatorial nature of joint action spaces, and complex information structures. (Kaiqing Zhang, Yang, and Başar 2019)

  • Consider using a hierarchical reinforcement learning approach for aggregated search tasks, consisting of a high-level source selection policy and a low-level item presentation policy, both trained with the DQN algorithm and receiving rewards from implicit user feedback. (Takanobu et al. 2019)

  • Use the Agent-by-agent Policy Optimization (A2PO) algorithm to improve sample efficiency and maintain monotonic improvement guarantees for each agent during training, while considering the impact of agent updating order and extending the theory of non-stationarity into the sequential update scheme. (Albrecht and Stone 2018)

  • Focus on developing and evaluating algorithms that enable efficient learning from limited data sources, particularly in scenarios where traditional supervised learning techniques may struggle due to insufficient data availability. (Allen-Zhu, Li, and Song 2018)

  • Carefully evaluate the impact of your prior knowledge on your predictions, particularly in situations where there might be a significant difference between the training and testing environments. By doing so, they can avoid potential pitfalls caused by overly reliant on past experiences that may no longer apply in the present situation. (P. Anderson et al. 2018)

  • Incorporate a “relational inductive bias” in your deep learning algorithms, which enables the system to make better decisions based on relationships between objects rather than just individual attributes. (Hamrick et al. 2018)

  • Use a search session Markov decision process (SSMDP) to model multi-step ranking problems in e-commerce applications, allowing for the optimization of long-term accumulative rewards through reinforcement learning techniques. (Yujing Hu et al. 2018)

  • Consider using value function learning instead of policy gradient methods for your reinforcement learning projects, as it tends to be more stable and sample efficient in cases where it applies. (J. Lim et al. 2018)

  • Utilise a scalable neural network design capable of handling DAG structures of varying shapes and sizes, and incorporating reinforcement learning techniques to effectively manage stochastic job arrivals, thereby significantly improving the efficiency of scheduling processes. (Hongzi Mao et al. 2018)

  • Develop a novel deep reinforcement learning (DRL) method for the SS-RTB problem, capable of effectively handling the environment changing problem through the use of a robust MDP at the hour-aggregation level, allowing for real-time bidding via capturing impression-level features and periodically controlling the bidding model according to real feedback from the environment. (Jun Zhao et al. 2018)

  • Utilise model-free reinforcement learning to develop agents capable of sequentially regulating bidding parameters in a highly non-stationary environment, thereby achieving near-optimal bidding strategies while avoiding excessive computational costs. (D. Wu et al. 2018)

  • Adopt a modular framework like RLgraph to effectively address the challenges associated with implementing, executing, and testing reinforcement learning tasks, thereby improving overall performance and efficiency. (“Proceedings of the 1st ACM SIGPLAN International Workshop on Machine Learning and Programming Languages” 2017)

  • Consider employing deep asynchronous stochastic Q-learning (DASQN) algorithms to address the challenge of handling large-scale real-world problems with high-level semantic information inputs, as demonstrated by the successful implementation of the LADDER agent in JDs real-time bidding (RTB) ad business.’ (Yu Wang et al. 2017)

  • Consider using competitive multi-agent environments trained with self-play to generate complex behaviors in agents, as these environments offer a natural curriculum and can produce behaviors that are more complex than the environment itself. (T. Bansal et al. 2017)

  • Consider using either multi-agent importance sampling or multi-agent fingerprints to effectively incorporate experience replay into multi-agent reinforcement learning, thus allowing for stable training of deep multi-agent value functions. (J. Foerster et al. 2017)

  • Consider utilizing “demonstration data” - i.e., data collected from previous control of the system - to significantly speed up the learning process in deep reinforcement learning models. (Hester et al. 2017)

  • Consider incorporating sequential social dilemmas into your models to better account for the temporally extended nature of real-world social dilemmas, as this can impact the emergence and stability of cooperation. (Leibo et al. 2017)

  • Integrate deep learning and reinforcement learning to achieve significant advancements across diverse scientific domains. (Chao Li et al. 2017)

  • Employ a two-phase approach to develop a decentralized multi-task multi-agent reinforcement learning (MT-MARL) system under partial observability, involving initial coordination in single-task MARL followed by distillation of specialized action-value networks into a generalized recurrent multi-task network. (Omidshafiei et al. 2017)

  • Utilize a Multiagent Bidirectionally-Coordinated Network (BiCNet) with a vectorized extension of actor-critic formulation to achieve efficient learning for intra-agent communication and coordination in multiagent systems. (P. Peng et al. 2017)

  • Consider employing robust adversarial reinforcement learning (RARL) techniques when working with deep neural networks in reinforcement learning contexts, as it helps improve training stability, enhances robustness to differences in training and test conditions, and outperforms baselines even without the presence of an adversary. (Pinto et al. 2017)

  • Consider using policy gradient iterations without Markovian assumptions, decomposing the problem into a composition of a Policy for Desires and trajectory planning with hard constraints, and introducing a hierarchical temporal abstraction called an “Option Graph” with a gating mechanism to reduce the variance of the gradient estimation and improve the efficiency of safe, multi-agent reinforcement learning for autonomous driving. (Shalev-Shwartz, Shammah, and Shashua 2016)

  • Consider implementing a dual-learning mechanism when working with neural machine translation (NMT) systems, as it enables automatic learning from unlabelled data through a dual-learning game, reducing the need for expensive human labelling and improving translation accuracy. (Y. Xia et al. 2016)

  • Consider integrating semantic and topological representations into your deep reinforcement learning models to improve performance in complex 3D environments. (Bhatti et al. 2016)

  • Utilize deep distributed recurrent Q-networks (DDRQN) to enable teams of agents to learn to solve communication-based coordination tasks without any pre-designed communication protocol. (J. N. Foerster et al. 2016)

  • Consider combining fictitious self-play with deep reinforcement learning to create a scalable end-to-end approach for learning approximate Nash equilibria in large-scale games of imperfect information, without requiring prior domain knowledge. (Heinrich and Silver 2016)

  • Utilise deep reinforcement learning with a curriculum learning scheme to tackle problems previously deemed intractable by most multi-agent reinforcement learning algorithms. (Houthooft et al. 2016)

  • Focus on developing end-to-end frameworks for task-oriented dialog systems using deep reinforcement learning techniques, which can address the credit assignment problem and process interdependence issues present in traditional pipelines. (Tiancheng Zhao and Eskenazi 2016)

  • Utilise a combination of deep reinforcement learning and a physics engine-integrated simulation framework like AI2-THOR to enable efficient, adaptable, and flexible learning for target-driven visual navigation tasks. (Yuke Zhu et al. 2016)

  • Consider implementing a dual-learning mechanism for machine translation, which enables an NMT system to automatically learn from unlabelled data through a dual-learning game, thereby reducing the need for expensive human labelling. (Gulcehre et al. 2015)

  • Manipulate the reward structure in multiagent systems to encourage desired behaviors, such as competition or collaboration, and observe the effects on the systems overall performance.’ (Tampuu et al. 2015)

  • Utilise Stochastic Value Gradient (SVG) methods when optimising stochastic policies in stochastic environments. These methods combine the benefits of both model-based and model-free approaches, reducing the impact of model errors and increasing overall efficiency. (Balduzzi and Ghifary 2015)

  • Be aware of the potential for overestimation in Q-learning algorithms, particularly when combining it with deep neural networks like in the DQN algorithm, and consider implementing methods like Double Q-learning to mitigate this issue and improve overall performance. (Hasselt, Guez, and Silver 2015)

  • Employ a deep reinforcement learning framework to jointly learn state representations and action policies using game rewards as feedback, enabling mapping of text descriptions into vector representations that capture the semantics of the game states. (Narasimhan, Kulkarni, and Barzilay 2015)

  • Leverage deep reinforcement learning and model compression techniques to train a single policy network that learns how to act in a set of distinct tasks by using the guidance of several expert teachers, enabling efficient multitask and transfer learning across various domains. (Parisotto, Ba, and Salakhutdinov 2015)

  • Utilise a “dueling” neural network architecture in your reinforcement learning models, which separates the estimation of state value and advantage functions, allowing for faster convergence and better performance in complex environments. (Ziyu Wang et al. 2015)

  • Consider the impact of non-stationarity in multi-agent environments when developing reinforcement learning algorithms, as traditional methods designed for single-agent domains may not perform effectively in these complex settings. (H. M. Schwartz 2014)

  • Use a convolutional neural network trained with a variant of Q-learning to enable successful control policies to be learned directly from high-dimensional sensory input, such as raw video data, in complex reinforcement learning environments. (V. Mnih et al. 2013)

  • Adopt a hierarchical Bayesian framework for individual learning under uncertainty, which addresses the limitations of traditional Bayesian and reinforcement learning models through efficient, interpretable, and flexible updates that accommodate inter-individual variability. (Mathys 2011)

  • Utilise a combination of reinforcement learning and deep learning approaches to create effective real-time Atari game playing agents, specifically by leveraging slow planning-based agents to provide training data for a deep-learning architecture capable of real-time play. (Y. Bengio 2009)

  • Consider using sequential Monte Carlo methods for dynamic systems, which involve a combination of importance sampling and resampling, rejection sampling, and Markov chain iterations, depending on the specific circumstances of the application. (NA?)

  • Consider implementing Iterated Guaranteed Safe Online Learning via Reachability (IGSOLR) in your studies, which involves modeling worst-case disturbances in a state-dependent manner, learning these models online, and periodically recomputing safe sets to achieve a balance between safety and performance. (NA?)

  • Utilise the Mean Square Projected Bellman Error (MSPBE) as an objective function for developing stochastic gradient-descent algorithms for temporal-difference learning with linear function approximation, as opposed to the previously used Mean Square Bellman Error (MSBE) or Norm of Expected TD Update (NEU). (NA?)

  • Adopt a merging technique that combines multiple matrix elements to accurately model complex systems, ensuring that the computational complexity remains manageable while providing reliable predictions. (NA?)

  • Carefully consider the balance between bias and overfitting in your machine learning models, taking into account factors like model complexity, data availability, and the nature of the task at hand. (NA?)

  • Focus on developing semi-supervised deep reinforcement learning approaches to effectively utilize the vast amounts of unlabelled data generated by smart cities, thereby improving the efficiency and effectiveness of smart city services. (NA?)

  • Utilise a scalable neural network design capable of handling DAGs of varying shapes and sizes, and incorporating reinforcement learning techniques to effectively manage unpredictable job arrival sequences. (NA?)

  • Consider using deep reinforcement learning techniques such as MolDQN for molecule optimization, as it combines domain knowledge of chemistry with advanced reinforcement learning strategies, ensures 100% chemical validity, operates without pre-training on any dataset, and enables multi-objective optimization. (NA?)

  • Explore the potential of machine learning techniques, particularly low-complexity Q-learning approaches, to optimize the operation of ultra-dense cellular IoT networks, given the challenges posed by the massive number of MTC devices, dynamic traffic patterns, and limited radio resources. (NA?)

Multi-Agent Systems

  • Use the Optimal Baseline (OB) technique to achieve minimal variance in multi-agent policy gradient (MAPG) estimators, thereby improving the stability and performance of multi-agent reinforcement learning (MARL) algorithms. (Kuba et al. 2021)

  • Utilize a bi-level learning hierarchy in order to simplify the process of role discovery in multi-agent learning tasks. (Tonghan Wang et al. 2020)

  • Consider using massive multiplayer online role-playing games (MMORPGs) as platforms for studying complex multiagent interactions, due to your ability to simulate large, persistent populations and support diverse behaviors. (Suarez et al. 2019)

  • Consider the interdependence between environment and co-player when developing curricula for multi-agent reinforcement learning systems, as doing so can lead to improved performance and robustness. (Noam Brown and Sandholm 2019)

  • Consider utilising the State Inference for value DEcomposition (SIDE) framework when dealing with partially observable problems in multi-agent reinforcement learning, as it enables simultaneous solution of optimal control and state inference issues. (S. Pan et al. 2018)

  • Employ a monotonic value function factorization technique, specifically QMIX, to effectively train decentralized policies in a centralized end-to-end fashion within multi-agent reinforcement learning contexts. (P. Peng et al. 2017)

  • Adopt an alternating optimization approach to solve complex problems involving multiple simultaneous learning agents, integrating imitation learning with unsupervised structure learning by taking turns to optimize for imitation policies while fixing a structured model, and vice versa. (H. M. Le et al. 2017)

  • Utilise a fully decentralised multi-agent reinforcement learning approach, leveraging a networked communication structure between agents, to effectively solve complex tasks in large scale environments. (Gruslys et al. 2017)

  • Avoid overfitting to the other agents policies in multiagent reinforcement learning (MARL) scenarios by employing a novel algorithm that computes approximate best responses to mixtures of policies generated using deep reinforcement learning and employs empirical game-theoretic analysis to compute meta-strategies for policy selection. (Lanctot et al. 2017)

  • Adapt actor-critic methods to consider the action policies of other agents in multi-agent domains, leading to successful learning of complex multi-agent coordination. (R. Lowe et al. 2017)

  • Study the emergence of grounded compositional language in multi-agent populations by creating a physically-situated multi-agent learning environment and developing learning methods that enable agents to communicate effectively and efficiently to achieve goals in various scenarios. (Mordatch and Abbeel 2017)

  • Employ a monotonic value function factorization technique, specifically QMIX, to effectively train decentralized policies in a centralized end-to-end fashion within multi-agent reinforcement learning contexts. (P. Peng et al. 2017)

  • Carefully consider and control for the coordination and heterogeneity levels of the environment when conducting multi-agent reinforcement learning experiments, as these factors significantly impact the performance of various algorithms. (Oliehoek and Amato 2016)

  • Adapt actor-critic methods to consider the action policies of other agents in multi-agent domains, leading to successful learning of policies requiring complex multi-agent coordination. (Abadi and Andersen 2016)

  • Explore developing and utilizing architectures capable of dealing with richer and more complex dynamics in multi-agent systems, as demonstrated by the superior performance of DRUQN and DLCQN compared to DQN in various competitive and cooperative experiments. (Abel et al. 2016)

  • Embrace deep neural networks to enable end-to-end learning of communication protocols in complex environments, utilizing techniques like Reinforced Inter-Agent Learning (RIAL) and Differentiable Inter-Agent Learning (DIAL) to optimize coordination among agents. (J. N. Foerster et al. 2016)

  • Consider using a neural model like CommNet, which learns communication among agents alongside your policies, to improve performance in collaborative tasks. (Sukhbaatar, Szlam, and Fergus 2016)

  • Consider the potential impact of “cheap talk” channels on the results of your study, and explore ways to identify and utilise these channels to enhance the quality of your findings. (Diederik P. Kingma and Ba 2014)

  • Consider extending existing Bayesian IRL approaches to handle situations where the reward depends on both state and action, and then compare these extended approaches to MIRL methods in various scenarios to determine your effectiveness. (X. Lin, Beling, and Cogill 2014)

  • Adopt a sequential decision making paradigm for studying negotiation, focusing on the negotiation process itself instead of solely on outcomes, and incorporating learning mechanisms to improve the performance of autonomous agents. (“Complex Automated Negotiations: Theories, Models, and Software Competitions” 2013)

  • Carefully consider the choice of input/output logic when studying norm change in multiagent systems, as different logics may yield distinct results and interpretations. (Boella and Torre 2007)

  • Carefully consider the role of cooperation in enabling the emergence of language, as demonstrated by the difference in performance between self-interested and prosocial agents in the negotiation environment. (C. V. Goldman, Allen, and Zilberstein 2006)

  • Carefully consider the choice of conceptual model, information sources, visibility types, model granularity, agent behavior assumptions, type of exchanged information, and trust/reputation reliability measure when developing computational trust and reputation models. (Sabater and Sierra 2005)

  • Utilise the concept of Influence-Based Multi-Agent Exploration’, specifically using either ‘Exploration via Information-Theoretic Influence’ (EITI) or ‘Exploration via Decision-Theoretic Influence’ (EDTI) techniques, to improve coordination and exploration efficiency in multi-agent systems.’ (Chalkiadakis and Boutilier 2003)

  • Consider adopting a decentralized Bayesian approach to coordinating multiple autonomous sensor platforms searching for a single non-evading target, as it allows for scalability, modularity, and real-time adaptability while maintaining accurate tracking of the target state probability distribution. (Bourgault, Furukawa, and Durrant-Whyte, n.d.)

  • Adopt the Markov game framework instead of the traditional Markov decision process (MDP) model for studying multi-agent reinforcement learning scenarios, as it enables consideration of multiple adaptive agents with interacting or competing goals. (NA?)

  • Carefully consider the trade-offs between homogeneous and heterogeneous team learning approaches in multi-agent systems, taking into account factors such as problem domain requirements, search space size, and potential benefits of agent specialization. (NA?)

  • Consider using an asymmetric learning model for multiagent reinforcement learning tasks, especially in semi-centralized multiagent systems, due to its unique equilibrium point value, faster equilibrium point evaluation, and potential for reduced space and computational requirements compared to symmetric models. (Kononen, n.d.)

Generative Models

Probabilistic Graphical Models

  • Utilize a joint semiparametric model for several potentially related multi-state processes, assuming a Markov structure for the transitions over time, and capturing the dependence between different processes using a joint prior distribution on the transition rates of each process. (Cremaschi et al. 2023)

  • Utilize a Gaussian variational approximation (GVA) for high-dimensional state space models, specifically by employing a dynamic factor model to reduce the complexity of the covariance matrix while maintaining the essential dependencies within the data. (Quiroz, Nott, and Kohn 2023)

  • Use a generalized class of mixtures of finite mixtures (MFMs) with a hyperparameter dependent on the number of components, rather than traditional MFMs with a fixed hyperparameter, to improve the performance of Bayesian non-parametric mixture models. (Frühwirth-Schnatter, Malsiner-Walli, and Grün 2021)

  • Utilize Bayesian networks to efficiently represent complex probabilistic relationships among multiple variables, enabling accurate inferences and predictions even in situations with limited available data. (Heckerman 2020)

  • Consider implementing a forced pruning learner algorithm for structure learning in Markov networks, which involves a combination of greedy edge deletion and rejection sampling techniques to optimize performance and reduce computational costs. (Abdelatty, Sahoo, and Roy 2018)

  • Utilize the probability product kernel, a novel kernel between distributions, to effectively combine the benefits of discriminative learning algorithms and kernel machines with generative modeling techniques. (Sunghwan Kim et al. 2015)

  • Adopt the framework of continuous time Bayesian networks (CTBNs) for modelling complex temporal processes, as it enables explicit representation of temporal dynamics and efficient querying of the distribution over specific event times, while allowing for flexible handling of irregularly spaced observations. (Nodelman, Shelton, and Koller 2013)

  • Utilize the Bayesian Structural EM algorithm for learning probabilistic models, specifically Bayesian networks, from incomplete data, as it offers improved accuracy compared to traditional approaches like the BIC score. (N. Friedman 2013)

  • Develop a three-phase algorithm for efficiently constructing Bayesian belief networks from databases, utilizing mutual information and conditional independence tests to minimize unnecessary calculations and improve accuracy. (M. Singh and Valtorta 2013)

  • Utilise loopy belief propagation (BP) as a powerful tool for approximate learning and inference in dependency parsing, due to its ability to effectively incorporate global constraints and higher-order features without compromising computational efficiency. (Sanchez-Vega et al. 2013)

  • Utilize integer programming, specifically the SCIP framework, to optimize the structure of Bayesian networks by maximizing log marginal likelihood (BDe score) through the use of cutting planes, which are effective in reducing the upper bound given by LP solutions and increasing the efficiency of the optimization process. (Cussens 2012)

  • Adopt a Bayesian probabilistic graphical model called Latent Truth Model (LTM) to automatically infer true records and source quality without any supervision, as it enables simultaneous determination of source quality and inferring underlying truth through iteration, supports multiple truths for the same entity, and allows for efficient and scalable linear complexity inference algorithm. (Bo Zhao et al. 2012)

  • Focus on developing algorithms for learning latent tree graphical models that are both consistent and computationally efficient, particularly for cases where not all variables are observed. (M. J. Choi et al. 2010)

  • Utilize a Bayesian hierarchical model when analyzing football data, but must be aware of potential overshrinkage issues and consider implementing a more complex mixture model to mitigate them. (Baio and Blangiardo 2010)

  • Consider using the bnlearn R package for learning the structure of Bayesian networks, as it offers multiple algorithms for handling both discrete and continuous variables, supports parallel computing for improved performance, and allows users to choose the best combination of learning algorithms and statistical criteria for your specific dataset. (Scutari 2009)

  • Utilize variational Bayesian learning techniques when dealing with complex graphical models, particularly those falling under the category of conjugate-exponential models, as they offer significant benefits such as efficient inference and robustness against overfitting. (Jian Zhang, Ghahramani, and Yang 2008)

  • Utilize marginal data augmentation techniques in your Bayesian analyses of multinomial probit models, as it provides significant improvements in computational efficiency and ease of interpretation compared to alternative approaches. (Imai and Dyk 2005)

  • Utilise the novel vine’ graphical model for dependent random variables, which extends traditional Markov trees and Bayesian belief nets by allowing for different types of conditional dependence, making it easier to incorporate expert knowledge and perform simulations.’ (Bedford and Cooke 2002)

  • Utilise the Naive Bayesian Classifier approach due to its ability to effectively estimate the probability of a certain event occurring based on prior knowledge and new evidence, despite making the simplifying assumption of class conditional independence. (Henderson 2002)

  • Focus on developing generally applicable “building blocks,” called idioms, which can be combined into larger Bayesian networks using simple combination rules and by leveraging recent advancements in modular and object-oriented Bayesian networks (OOBNs). (NEIL, FENTON, and NIELSON 2000)

  • Carefully select and justify your choice of covariance function when implementing Gaussian processes for regression and classification tasks, taking into account factors such as smoothness, relevance determination, and potential interactions among input dimensions. (S. A. Goldman and Sloan 1995)

  • Consider using Bayesian inference in models for density estimation using mixtures of Dirichlet processes, as they provide natural settings for density estimation and allow for direct inference on a variety of practical issues, including problems of local versus global smoothing, uncertainty about density estimates, assessment of modality, and the inference on the numbers of components. (Escobar and West 1995)

  • Consider using Bayesian inference in models for density estimation using mixtures of Dirichlet processes, as they provide natural settings for density estimation and allow for direct inference on a variety of practical issues, including problems of local versus global smoothing, uncertainty about density estimates, assessment of modality, and the inference on the numbers of components. (Escobar and West 1995)

  • Utilize a Bayesian approach to infer the probability of identity between two objects, using an identity criterion to relate propositional observations to identity sentences, and leveraging appearance probabilities to compute this probability in terms of observable quantities. (Ingemar J. Cox 1993)

  • Focus on developing methods that leverage the power of Markov blankets to efficiently learn Bayesian networks from data, thereby producing more compact and interpretable causal models. (Chow and Liu 1968)

  • Utilize the proposed novel exact algorithm for structure discovery in Bayesian networks, which offers significant improvements in efficiency compared to previous algorithms, allowing for accurate predictions and inferences in complex scenarios. (PEARSON 1905)

  • Utilize the property of likelihood equivalence in your scoring metric when learning Bayesian networks, as it allows for accurate inference of causality from observational data. (NA?)

  • Carefully consider your choice of mean and covariance functions when working with Gaussian processes, as they play a crucial role in determining the performance and accuracy of the model. (NA?)

  • Focus on improving the computational efficiency of existing methods for weakening the attribute independence assumption in naive Bayes classifiers, while still maintaining comparable prediction accuracy. (NA?)

  • Consider utilizing Markov Logic Networks (MLNs) as a powerful tool for combining first-order logic and probabilistic graphical models in a single representation, thereby providing a compact language to specify large Markov networks and the ability to incorporate a wide range of domain knowledge. (NA?)

  • Consider extending the Hierarchical Dirichlet Process Hidden Markov Model (HDP-HMM) to include explicit-duration semi-Markov modeling, which allows for more accurate representation of non-geometric state durations and improved interpretability of results. (NA?)

  • Consider using advanced computational methods, such as Bayesian learning and causal Bayesian networks, to better understand and analyze the complex ways in which children learn and process information. (NA?)

  • Carefully consider the choice of probabilistic network type (such as discrete Bayesian networks, conditional linear Gaussian Bayesian networks, discrete influence diagrams, conditional linear-quadratic Gaussian influence diagrams, limited-memory influence diagrams, or object-oriented probabilistic networks) depending on your specific application requirements, taking into account factors like the nature of the variables involved (continuous or discrete), the presence of decision variables and utility functions, the availability of historical data, and the desired level (NA?)

  • Utilise loopy belief propagation (BP) as a powerful tool for approximate learning and inference in dependency parsing, due to its ability to effectively incorporate global constraints and higher-order features without compromising computational efficiency. (NA?)

  • Consider using Bayesian networks when studying complex systems involving multiple interacting factors, as they allow for efficient and accurate analysis of uncertain relationships among variables. (NA?)

  • Utilise the novel vine’ graphical model for dependent random variables, which extends traditional Markov trees and Bayesian belief nets by allowing for different types of conditional dependence, making it easier to specify multivariate distributions and perform simulations.’ (NA?)

  • Consider integrating Bayesian Networks (BNs) with other modeling frameworks and tools to enhance the applicability and reliability of your findings across various domains, while remaining mindful of potential limitations and challenges arising from increased complexity. (NA?)

  • Utilise the walk-sum’ formulation for computation of means, variances and correlations as sums over certain sets of weighted walks in a graph, which applies to a wide class of Gauss-Markov models called ‘walk-summable’. (NA?)

Hidden Markov Models (Hmm)

  • Focus on building joint probabilistic models of time series data, which involves predicting future outcomes based on previous observations and detecting changes in behavior when the probability of observing certain patterns becomes significantly smaller. (Limoyo, Ablett, and Kelly 2022)

  • Utilize the newly proposed algorithms for exact Bayesian inference in the sparse normal sequence model, which significantly improve upon the current limitations of existing algorithms in terms of numerical instability and computational efficiency. (Erven and Szabó 2021)

  • Utilise Dynamic Bayesian Networks (DBNs) for modelling sequential data due to your ability to generalise Hidden Markov Models (HMMs) and Kalman Filter Models (KFMs) by allowing factorized representation of state spaces and arbitrary probability distributions respectively. (Shiguihara, Lopes, and Mauricio 2021)

  • Utilise SrVARM, a state-regularised autoregressive model, to effectively capture the dynamics of transitions between a finite set of hidden states and the state-dependent DAG-structured inter-variable dependencies within multi-dimensional time series data. (T.-Y. Hsieh et al. 2021)

  • Consider using marginalization techniques instead of simulating discrete latent states in Bayesian population models, as it significantly speeds up calculations while maintaining similar levels of accuracy. (Yackulic et al. 2020)

  • Consider using hierarchical Hidden Markov Models (HHMMs) to analyze multi-scale time series data, particularly in economics, as these models allow for the integration of variables observed at varying temporal resolutions, providing a more comprehensive understanding of stock market dynamics. (Noè et al. 2019)

  • Utilize hierarchical modeling techniques in nonparametric settings to effectively manage the high degree of freedom associated with nonparametric models, thereby enabling the creation of more complex and flexible probabilistic structures. (Teh and Jordan 2010)

  • Consider using variable length Markov chains (vlmc) for modeling stationary categorical processes due to your ability to overcome limitations of traditional fixed-order Markov chains, such as limited structural richness and the curse of dimensionality. (“Variable Length Markov Chains: Methodology, Computing and Software” 2002)

  • Utilise advanced statistical techniques like Hidden Markov Models (HMMs) and Bayesian Networks when dealing with sequential data, especially where the underlying processes generating the data may be unseen or “hidden”. These models offer powerful tools for making predictions and drawing inferences even when the data is incomplete or uncertain, thanks to your ability to capture the inherent temporal dependencies and conditional independence structures within the data. (GHAHRAMANI 2001)

  • Consider using distantly-labeled data to set model parameters, as it significantly improves extraction accuracy in hidden Markov models for information extraction tasks. (NA?)

  • Consider multiple machine learning methods for sequential data analysis, including sliding window methods, recurrent sliding windows, hidden Markov models, conditional random fields, and graph transformer networks, while taking into account factors like loss functions, feature selection, and computational efficiency. (NA?)

  • Use machine learning algorithms like hidden Markov models to analyze complex longitudinal datasets in order to reveal novel, unbiased phenotypes of atopy that can improve our understanding of the etiology of asthma. (NA?)

  • Focus on selecting appropriate sensors, acquiring quality data, evaluating relevant features, and employing suitable machine learning techniques, particularly Hidden Markov Models, for accurately classifying human physical activity from on-body accelerometers. (NA?)

  • Consider reframing your stochastic optimal control problems as Kullback-Leibler (KL) minimization problems, allowing them to leverage powerful graphical model inference techniques for efficient and accurate solutions. (NA?)

  • Consider utilizing hybrid generative/discriminative models, specifically combining Hidden Markov Models (HMM) with Artificial Neural Networks (ANN) or Support Vector Machines (SVM), for improved activity recognition in home environments using binary sensors. (NA?)

Conditional Random Fields (Crf)

  • Utilise Maximum Margin Markov (M^3) networks, which integrate the benefits of kernel-based approaches and probabilistic graphical models, to improve the performance of classification tasks involving sequential, spatial, or structured data. (Nowak-Vila, Bach, and Rudi 2020)

  • Consider using Gibbs sampling, a simple Monte Carlo method, to perform approximate inference in factored probabilistic models, particularly in situations where traditional dynamic programming approaches fail to capture long-distance structural relationships present in complex data sets. (Finkel, Grenager, and Manning 2005)

  • Consider using dynamic conditional random fields (DCRFs) for modeling complex interactions between labels in sequence data, as they offer improved performance compared to traditional linear-chain CRFs while allowing for rich, overlapping feature sets. (NA?)

  • Combine Conditional Random Fields (CRFs) and a variation of AutoSlog (AS) for optimal performance in identifying sources of opinions, emotions, and sentiments in text. (NA?)

Restricted Boltzmann Machines (Rbm)

  • Utilise deep structured energy based models (DSEBMs) when dealing with anomaly detection problems. These models offer significant advantages because they enable direct modelling of the data distribution using deep architectures, allowing for adaptation to different types of data structures like static, sequential, and spatial data. Furthermore, these models can be trained efficiently using score matching techniques, avoiding the need for complex sampling methods. Finally, the authors suggest two decision criteria - the energy score and the reconstruction (Schölkopf et al. 2001)

  • Utilize Annealed Importance Sampling (AIS) to accurately estimate ratios of partition functions in Restricted Boltzmann Machines (RBMs) and Deep Belief Networks (DBNs), thereby enabling more precise model selection and complexity control. (NA?)

  • Consider using a low-rank approximation to the interaction tensor in your spatial transformation models, which involves a sum of factors’ each of which is a three-way outer-product. This approximation allows for efficient learning of transformations between larger image patches and enables the model to learn optimal filter pairs for efficiently representing transformations.’ (NA?)

Variational Inference

  • Carefully distinguish between infinite volume results and finite volume ones, especially when dealing with systems close to a first order phase transition, as the presence of long-range order near boundaries can lead to non-uniform decay of correlation functions and a global slowdown of dynamics. (R. Han and Yang 2024)

  • Avoid relying solely on a single series from the Gibbs Sampler for your analyses, as it could lead to false precision and misleading results. Instead, they recommend using multiple independent sequences to ensure accurate and reliable conclusions. (Craiu, Gong, and Meng 2023)

  • Utilize score-based generative models to effectively identify anomalous nodes within attributed networks, by analyzing the discrepancies between the original and reconstructed ego-graphs. (Gavrilev and Burnaev 2023)

  • Prioritize running longer Markov Chain Monte Carlo (MCMC) simulations rather than relying solely on shorter ones, as longer runs provide better estimates and reduce the risk of underestimating the true variance. (Margossian and Gelman 2023)

  • Utilize a localized version of the potential scale reduction factor (()) to improve the diagnostic performance of Markov Chain Monte Carlo (MCMC) convergence, allowing for better identification of convergence issues across various quantiles of the target distribution. (Moins et al. 2023)

  • Utilize the branching process representation of the Bayesian Context Trees (BCT) model prior and posterior to develop efficient algorithms for model selection, estimation, and prediction tasks. (Papageorgiou and Kontoyiannis 2023)

  • Consider using marginally augmented variational Bayes (MAVB) to improve your initial variational approximation in order to achieve better accuracy and scalability in analyzing non-nested binomial hierarchical models. (Goplerud 2022)

  • Consider using Gaussian orthogonal latent factor processes for modeling and predicting large correlated datasets, as it offers significant computational efficiency through its ability to decompose the likelihood function into a product of densities at orthogonal components with lower-dimensional inputs. (M. Gu and Li 2022)

  • Consider using Expectation Propagation (EP) rather than Laplaces method for approximate Bayesian inference in binary Gaussian processes classification because EP offers better predictive performance and more accurate marginal likelihood estimates compared to Laplace’s method.’ (Shapovalova, Heskes, and Dijkstra 2022)

  • Leverage the Givens representation to enable efficient and accurate sampling from distributions over the Stiefel manifold, thereby enabling the development of advanced statistical models with orthogonal matrix parameters. (Pourzanjani et al. 2021)

  • Replace traditional trace plots with rank plots from multiple chains, and utilize rank-based diagnostics and quantile-based local efficiency measures to improve the monitoring of convergence in Markov Chain Monte Carlo (MCMC) algorithms. (Vehtari et al. 2021)

  • Adopt a unified stochastic gradient approach to Bayesian optimal experimental design (BOED) that enables simultaneous optimization of both variational and design parameters, leading to increased efficiency and scalability compared to traditional two-stage frameworks. (A. Foster et al. 2019)

  • Utilize the Pareto Smoothed Importance Sampling (PSIS) diagnostic tool to assess the quality of your variational inference (VI) approximation to the full posterior distribution, and subsequently decide whether to improve the VI approximation or switch to exact sampling techniques like Markov Chain Monte Carlo (MCMC). (Yuling Yao et al. 2018a)

  • Utilise a variational algorithm for achieving approximate maximum likelihood inference when dealing with multivariate non-Gaussian observations, particularly in fields like ecology or genomics. (Chiquet, Mariadassou, and Robin 2017)

  • Consider using variational inference as an alternative to traditional MCMC methods for approximating probability densities, particularly in cases where data sets are large or models are highly complex, due to its potential for increased efficiency and scalability. (Blei, Kucukelbir, and McAuliffe 2017)

  • Utilize Bayesian filtering techniques, specifically particle filters, for accurate online estimation in complex, nonlinear, non-Gaussian, and non-stationary systems. (Giron-Sierra 2017)

  • Utilize Generalized Empirical Likelihood (GEL) methods as a diagnostic tool to identify deficiencies in deep generative models (DGMs), such as mode dropping and mode imbalance, through specifying appropriate moment conditions. (Esteban, Hyland, and Rätsch 2017)

  • Develop a novel approach to data synthesis that involves setting up a 3-way competition among the synthesizer, target, and discriminator networks, ensuring that the synthesizer generates composite images that can fool both the target and discriminator networks, ultimately leading to efficient, task-aware, and realistic synthetic data generation. (Khoreva et al. 2017)

  • Utilize automatic differentiation variational inference (ADVI) to enable efficient cycling through the probabilistic modeling process, allowing for rapid exploration and refinement of complex models without needing to manually derive algorithms for each iteration. (Abadi et al. 2016)

  • Utilize the GPflow library, which leverages TensorFlow for efficient computations, employs variational inference for accuracy, offers automatic differentiation for conciseness, and enables GPU acceleration for faster processing. (G. Matthews et al. 2016)

  • Utilize a Bayesian approach to mixture modelling based on Student-\(t\) distributions rather than Gaussian distributions, as the former provides greater robustness against outliers and leads to more accurate estimations of the number of components in the mixture. (Chong Wang and Blei 2015)

  • Carefully analyze the bias and mixing time of asynchronous Gibbs sampling algorithms to ensure accurate and efficient results. (Mania et al. 2015)

  • Utilise normalising flows to create more accurate and complex posterior approximations in variational inference, leading to improvements in performance and applicability. (Dinh, Krueger, and Bengio 2014)

  • Utilize the Metropolis-Hastings Walker (MHW) sampler, which combines the Metropolis-Hastings algorithm and the alias method, to achieve significant improvements in computational efficiency for various probabilistic models, including latent Dirichlet allocation, Poisson Dirichlet processes, and hierarchical Dirichlet processes. (A. Q. Li et al. 2014)

  • Consider utilizing Bayesian filtering and smoothing techniques when dealing with complex data sets, as they provide a robust and flexible framework for incorporating prior knowledge and updating beliefs in light of new evidence. (Särkkä 2013)

  • Use the Macdonald processes, a family of probability distributions on sequences of partitions, to evaluate averages for a wide variety of observables, leading to a Fredholm determinant representation of a \(q\)-Laplace transform of the distribution of the last part of a Macdonald-random partition. (Borodin and Corwin 2013)

  • Utilise a robust and scalable Gaussian process regression (GPR) model via variational learning to enable the application of Gaussian processes to a wide range of real data, which are often large-scale and contaminated by outliers. (Hensman, Fusi, and Lawrence 2013)

  • Utilize stochastic optimization techniques, specifically stochastic gradient ascent, to efficiently optimize the Evidence Lower Bound (ELBO) in variational inference problems involving large or streaming datasets. (M. Hoffman et al. 2012)

  • Carefully evaluate the potential impact of compactness and parameter learning biases in variational expectation maximization (vEM) algorithms, particularly in time-series models, as these issues can significantly affect the accuracy and reliability of inferences. (Turner and Sahani 2011)

  • Carefully choose the number of simulations based on your inferential goals, and monitor convergence by comparing both within-chain and between-chain variability. (Hobert 2011)

  • Focus on analyzing the error estimates when approximating random walks with Brownian motion, especially for mean zero, finite variance walks, since the central limit theorem alone does not provide sufficient precision. (Lawler and Limic 2010)

  • Consider implementing slice sampling methods, which can adaptively change the scale of changes made and avoid problems arising from varying scales across distributions, leading to improved sampling efficiency compared to traditional Gibbs sampling and Metropolis methods. (Neal 2003)

  • Utilise multiple independent sequences with starting points sampled from an overdispersed distribution to achieve accurate and reliable inferences from iterative simulation. (X.-L. Meng 1994)

  • Utilise Bayesian inference through Monte Carlo simulation when dealing with mixture distributions, as it allows for accurate predictions and eliminates the need to decide on a correct number of components. (“Maximum Entropy and Bayesian Methods” 1992)

  • Consider utilizing Bayesian methods, specifically Gibbs sampling, for analyzing complex normal-mixture models in classification and discrimination tasks, as they offer significant advantages over traditional techniques, including computational efficiency, flexibility, and the ability to handle multiple normal-mixture components with varying covariance structures. (M. Lavine and West 1992)

  • Consider utilizing the free and open source C++ library libDAI for implementing various exact and approximate inference methods in probabilistic graphical models, due to its modular design, wide range of inference algorithms, and ease of implementation. (NA?)

  • Consider using the Student-\(t\) mixture model instead of the traditional Gaussian mixture model for density estimation, clustering, and model selection tasks because it provides greater robustness against outliers and allows for better handling of complex data structures. (NA?)

  • Consider incorporating heteroscedastic Gaussian processes into your statistical analyses, particularly when dealing with input-dependent noise rates, as this approach offers improved accuracy and computational efficiency compared to traditional methods. (NA?)

  • Adopt a direct importance estimation method for covariate shift adaptation, rather than relying on separate density estimation, because it allows for better model selection and improved prediction performance. (NA?)

  • Consider using Variational Bayesian Inference instead of Expectation Maximization (EM) algorithm for better results in complex statistical signal processing problems. (NA?)

  • Consider using the variational Gaussian approximation when dealing with models involving Gaussian priors and factorising likelihoods, as it reduces the number of variational parameters needed to optimize from O(N^2) to O(N), making it a more efficient choice for these types of models. (NA?)

  • Utilize stochastic variational inference, a scalable algorithm for approximating posterior distributions, to effectively analyze large datasets in various areas of science. (NA?)

  • Utilise the basic marginal likelihood identity (BMI) when computing the marginal likelihood of your data. This involves expressing the marginal likelihood as the ratio of the product of the sampling density and the prior to the posterior density of the parameter. By doing so, they avoid issues associated with previous approaches like instability due to infinite variances or the need for complex tuning functions. Furthermore, the authors suggest selecting a high density point for increased estimation accuracy. (NA?)

Graph Theory And Graph Neural Networks

  • Consider employing Deep Graph Neural Networks (GNNs) to tackle the challenging problem of new node prediction’ in graph mining, which involves predicting all links from a newly introduced, isolated node without prior link patterns.’ (Zanardini and Serrano 2024)

  • Consider using a hierarchical stochastic block model (HSBM) for community detection in multiplex networks, as it enables the modeling of varying communities across different network layers and effectively borrows information across layers for improved estimation. (A. Amini, Paez, and Lin 2024)

  • Consider using a combination of Graph AutoEncoder (GAE) and Graph Contrastive Learning (GCL) to improve the accuracy of group-level graph anomaly detection (Gr-GAD) by effectively capturing long-range inconsistencies and utilizing topology pattern information. (Ai et al. 2023)

  • Consider developing specialised techniques to fully utilise the rich semantics of directed multigraph data models for illicit account detection on cryptocurrency transaction networks. (Z. Ding et al. 2023)

  • Consider implementing a dynamic relation-attentive graph neural network (DRAG) for improved fraud detection, as it effectively addresses the challenge of heterophily in graphs through its ability to dynamically adjust attention coefficients for each node. (Heehyeon Kim, Choi, and Whang 2023)

  • Consider using temporal motifs as a valuable tool for analyzing financial transaction networks, as demonstrated by your successful application in fraud detection, link prediction, and node classification across three distinct financial networks: Mercari, JPMC, and Venmo. (Penghang Liu et al. 2023)

  • Develop a unified framework for understanding graph prompt learning, focusing on prompt tokens, token structures, and insertion patterns in the graph domain. (Xiangguo Sun et al. 2023)

  • Focus on developing a Spatial-Temporal-Aware Graph Transformer (STA-GT) model for transaction fraud detection, which combines a heterogeneous graph neural network with a temporal encoding strategy to effectively capture spatial-temporal information, and also includes a transformer module to learn local and global information. (Y. Tian et al. 2023)

  • Develop a novel open-set Graph Anomaly Detection (GAD) approach called normal structure regularization (NSReg) to leverage the rich normal graph structure embedded in the labelled nodes, which will help train an anomaly-discriminative supervised graph anomaly detector while preventing overfitting to the seen anomalies. (Qizhou Wang et al. 2023)

  • Adopt a “Pre-train, Prompt, and Predict” strategy when working with graph self-supervised learning methods, specifically using a strong and universal pre-training task called SGL that combines generative and contrastive self-supervised graph learning, and a novel verbalizer-free prompting function to unify the objectives of pre-text and downstream tasks. (Yun Zhu, Guo, and Tang 2023)

  • Utilize Bayesian nonparametric stochastic blockmodels as priors on the graph to facilitate the propagation of uncertainty in graph estimation to large-scale structure learning, thereby improving the effectiveness of information retrieval and interpretability. (Boom, Iorio, and Beskos 2023)

  • Consider utilizing the STRAND R package for its ability to simplify the implementation of generative network models in analyzing animal social network data, thereby addressing the current lack of easily accessible software options. (Ross, McElreath, and Redhead 2023)

  • Consider employing a generative adversarial network framework called “Adversarial Camouflage Detector” (ACD) to improve fraud detection by enhancing the ability to identify and remove camouflage in gang crime patterns, thereby improving the accuracy of spatiotemporal graph neural network models. (Lewen Wang et al. 2023)

  • Focus on developing individual calibration frameworks like CaliRare to improve the reliability of rare category analysis in graph-structured data by addressing the challenges of uncertainty and rarity through node-level uncertainty quantification algorithms and generalized distribution-based metrics like Expected Individual Calibration Error (EICE). (L. Wu et al. 2023)

  • Consider using a combination of structural and attribute information along with community detection techniques when developing models for group-based fraud detection on attributed bipartite graphs. (Jianke Yu et al. 2023)

  • Leverage both temporal and heterogeneous information to learn evolving node embeddings in continuous-time dynamic networks, specifically for the purpose of detecting anomalous behaviors in social networks. (Yilin Li et al. 2023)

  • Carefully consider various types of graph neural networks (GNNs) and interpretability methods, such as gradient-, perturbation-, and decomposition-based, surrogate models, and counterfactuals, along with advanced GNNs employed for several downstream applications, and evaluate them via a diverse set of metrics, efficiency results, interactive approaches, and novel ground truth, compared to earlier works. (Mobaraki and Khan 2023)

  • Strive to generate explanations that are both faithful to the GNN model and aware of the data, ensuring that the explanations are valid, meaningful, and trustworthy. (G. Lv and Chen 2023)

  • Strive to create a heterogeneity-agnostic multi-level explainer for deep graph networks (DGNs) that provides both topological- and feature- level explanations, while preserving the connectivity of the output subgraph. (G. Lv, Zhang, and Chen 2023)

  • Consider using transformer-based generative models for graphs, specifically for molecular graph generation, due to your ability to effectively capture the underlying structural properties of the training data and produce novel and realistic molecules. (Mazuz et al. 2023)

  • Consider using a dual relational graph attention network (DualGAT) approach when conducting event detection studies. This involves combining syntactic and semantic relations within a graph structure, allowing for improved accuracy in detecting events within text data. (Peng Li, Mi, and Mi 2022)

  • Consider using the Ray framework for efficient and scalable distributed training of graph neural network-based knowledge graph embedding models, achieving significant speedups in both data preprocessing and model training without compromising evaluation metrics. (N. Sheikh et al. 2022)

  • Incorporate domain-awareness into your bot detection systems using multi-relational graph neural networks, and employ federated learning to share data across different social networks while maintaining data privacy. (H. Peng et al. 2022)

  • Consider using INGREX, an interactive explanation framework for Graph Neural Networks, to enhance understanding of model predictions through dynamic and customizable visualizations. (Bui et al. 2022)

  • Leverage the combination of Benfords law and dense subgraph discovery to effectively detect anomalous subgraphs in financial and transaction networks. (Tianyi Chen and Tsourakakis 2022)

  • Consider using the Tree Movers Distance (TMD) as a pseudometric for attributed graphs, as it effectively bridges graph metrics and the stability of graph neural networks through a hierarchical optimal transport problem that accounts for both local attribute distribution and computation tree distribution.’ (C.-Y. Chuang and Jegelka 2022)

  • Employ the “pre-train, prompt, fine-tune” strategy for molecular representation learning, utilizing a motif prompting function to enhance generalization to various downstream tasks while avoiding the need for extensive pre-training objective engineering. (Diao et al. 2022)

  • Consider incorporating graph kernels into the message passing process of graph neural networks (GNNs) to enhance your performance and interpretability. (A. Feng et al. 2022)

  • Consider using the DGraph dataset for graph anomaly detection (GAD) research, as it addresses the limitations of current GAD datasets by providing a larger scale, incorporating temporal dynamics, and preserving background nodes, thus enabling deeper exploration of anomalous nodes. (X. Huang et al. 2022)

  • Consider using a neural sparsification framework like SparseGAD to improve the robustness of graph-based anomaly detection by optimizing the sparsification of graphs and collaboratively learning node representations in consideration of homophily and heterophily. (Kay Liu et al. 2022)

  • Consider using a deep graph learning approach, specifically the GLAD (Graph Learning for Anomaly Detection) model, to identify anomalous citations within citation networks. This involves incorporating text semantic mining to network representation learning via graph neural networks, allowing for the consideration of both node attributes and link attributes. Additionally, the CPU (Citation Purpose) algorithm can be employed to determine the purpose of citations based on citation texts. By doing so, researchers can better (Jiaying Liu et al. 2022)

  • Adopt a unified Graph-based sEmatic structure mining framework with ConTRAStive Learning (GETRAL) for evidence-based fake news detection. This involves modelling claims and evidences as graph-structured data, capturing long-distance semantic dependencies through neighbourhood propagation, reducing information redundancy using graph structure learning, feeding fine-grained semantic representations into a downstream claim-evidence interaction module, and applying supervised contrastive learning alongside adversarial (Junfei Wu et al. 2022)

  • Analyze the behavior of individual GNN neurons to gain insights into the interpretability of the entire model, enabling a high-level understanding of its functionality. (Xuanyuan et al. 2022)

  • Adopt a structure-aware approach when analyzing graph neural network (GNN) explanations, as opposed to relying solely on the traditional Shapley value method, which does not account for graph structures. (Shichang Zhang et al. 2022)

  • Develop trustworthy graph neural networks (GNNs) by incorporating essential aspects of trustworthiness, such as robustness, explainability, privacy, fairness, accountability, well-being, and other trust-oriented features, while taking into consideration the unique properties of graph data compared to traditional Euclidean space data. (He Zhang et al. 2022)

  • Utilise unsupervised model selection when working with graph neural networks (GNNs) for graph-level anomaly detection, as this allows for better identification of anomalous graphs in a collection of graphs. (L. Zhao et al. 2022)

  • Employ a Bayesian topological learning method to classify simulated actin filament networks, allowing for the prediction of cytoskeleton structural properties based on the number of cross-linking proteins in the network. (Maroulas, Micucci, and Nasrin 2022)

  • Utilise the proposed DynAnom framework for efficient anomaly detection in dynamic weighted-graphs, which enables accurate identification of both node and graph-level anomalies through effective localisation of periods when specific nodes undergo significant contextual changes. (X. Guo, Zhou, and Skiena 2022)

  • Adopt the ROLAND framework for dynamic graph analysis, which enables easy repurposing of static GNNs for dynamic graphs, offers a live-update evaluation setting that better mirrors real-world scenarios, and proposes a scalable and efficient training approach for dynamic GNNs via incremental training and meta-learning. (J. You, Du, and Leskovec 2022)

  • Employ a heterogeneous linguistics graph (HLG) to enhance Chinese pre-trained language models by integrating linguistics knowledge, leading to improved performance across six natural language processing tasks using ten benchmark datasets. (Yanzeng Li et al. 2022)

  • Utilise active learning techniques, particularly those incorporating graph convolutional networks (GCNs), to effectively manage imbalanced data distributions in complex systems. (L. Cui et al. 2022)

  • Consider implementing a Risk Graph Structure Learning (RGSL) model for prohibited item detection, which involves generating risk-aware item representations and searching risk-relevant pairs for structure learning iteratively, leading to significant improvements over traditional methods. (Y. Ji et al. 2022)

  • Utilise a combination of Graph Deviation Networks (GDN) and cross-network meta-learning algorithms to facilitate accurate anomaly detection in networks with limited labeled data. (“Proceedings of the Web Conference 2021” 2021)

  • Aim to strike a balance between structural and temporal information when attempting to detect phishing activities in blockchain systems. (Jinyin Chen et al. 2021)

  • Consider using the “Seen” method to enhance the explanation quality of graph neural network (GNN) model outputs by aggregating auxiliary explanations from important neighboring nodes, leading to improved explanation accuracy of up to 12.71%. (Hyeoncheol Cho, Oh, and Jeon 2021)

  • Consider distinguishing neighbours with different labels instead of excluding them when working with Graph Neural Networks (GNNs) for Graph Fraud Detection (GFD) tasks. (Hyunsoo Cho, Seol, and Lee 2021)

  • Focus on developing end-to-end learning frameworks that effectively combine multi-level social context and temporal information when attempting to detect fake news. (Jian Cui et al. 2021)

  • Use the CO-SNE algorithm for dimensionality reduction and visualization of hyperbolic data, as it effectively preserves both the global hierarchy and local similarity of high-dimensional hyperbolic embeddings in a low-dimensional hyperbolic space. (Yunhui Guo, Guo, and Yu 2021)

  • Extend existing explainability methods for Convolutional Neural Networks (CNNs) like Local Interpretable Model-agnostic Explanations (LIME), Gradient-Based Saliency Maps, and Gradient-weighted Class Activation Mapping (Grad-CAM) to Graph Neural Networks (GNNs) to effectively identify crucial edges in input graphs influencing GNN decision making. (Kasanishi, Wang, and Yamasaki 2021)

  • Focus on optimizing the computational procedures in graph convolutional networks (GCNs) for handling point clouds by reducing redundancy and exploiting the smooth propagation of local geometric structure information across GCNs. (Xiang Li et al. 2021)

  • Consider employing a Metapaths-guided Neighbors-aggregated Heterogeneous Graph Neural Network (MHN) model for heterogeneous graph embedding learning, which addresses limitations of previous models by incorporating node base embedding by attributes transformation, aggregation within one metapath, and aggregation among metapaths. (B. Lin et al. 2021)

  • Consider incorporating graph-based features derived from random walks in your analysis, as they can provide valuable information about the proximity of nodes to illicit activities in complex networks like the Bitcoin transaction network. (Oliveira et al. 2021)

  • Employ an adversarial active learning approach to develop a heterogeneous graph neural network for fake news detection, which uses a hierarchical attention mechanism to perform node representation learning in the heterogeneous information network, thereby enhancing learning performance even when faced with limited labeled data. (Yuxiang Ren et al. 2021)

  • Utilize a multi-purpose interpretation framework that involves acquiring a mask indicating topology perturbations of input graphs to effectively preserve, promote, or attack Graph Neural Network (GNN)s predictions.’ (Yi Sun et al. 2021)

  • Use the variance of payoff functions as an indicator to construct sparse coordination graphs, as it provides a reliable measure of the influence of one agents action on another’s expected utility, allowing for efficient and accurate decision making in multi-agent systems.’ (Tonghan Wang et al. 2021)

  • Utilise a probing framework to quantify the amount of meaningful information captured in graph representations, allowing them to better understand the inductive biases of graph-based models. (Feder et al. 2021)

  • Focus on developing asynchronous continuous time dynamic graph algorithms for real-time temporal graph embedding, as opposed to traditional graph models that execute two serial operations - graph querying and model inference. (Xuhong Wang et al. 2021)

  • Employ a combination of heterogeneous graph modeling and self-supervised learning techniques to effectively identify prohibited items in complex online marketplace environments. (Y. Ji, Shi, and Wang 2021)

  • Consider using GNNExplainer and GraphLIME to identify the optimal trigger injection positions for implementing effective and undetectable backdoor attacks on GNNs for graph classification and node classification tasks. (Jing Xu, Xue, and Picek 2021)

  • Use a causal graphical model to analyze the prediction generation process of Graph Convolutional Networks (GCNs) and estimate the causal effect of the local structure on GCN prediction, allowing them to adjust your model accordingly. (F. Feng et al. 2021)

  • Use a multi-scale temporal graph neural network framework like METRO to effectively capture dynamic and cross-scale variable correlations in multivariate time series forecasting tasks. (Y. Cui et al. 2021)

  • Consider using the LargeEA framework to align entities between large-scale knowledge graphs, as it addresses scalability issues through a combination of structure and name channels, allowing for more accurate and efficient entity alignment. (Congcong Ge et al. 2021)

  • Use a Reciprocal SpatioTemporal (REST) framework to improve spatiotemporal predictions by coupling spatial dependency inference using Edge Inference Networks (EINs) with Graph Convolutional Networks (GCNs) for temporal pattern modeling, resulting in improved prediction accuracy and discovery of meaningful spatial dependencies beyond predefined graphical structures. (H. Lin et al. 2021)

  • Consider adopting the KE-GCN framework, which allows for the simultaneous updating of both entity and relation embeddings within a graph convolutional network, leading to improved performance across multiple tasks. (D. Yu et al. 2021)

  • Utilise a neural-symbolic approach when dealing with complex tasks requiring both deep learning and logical reasoning, allowing for mutual reinforcement between these two processes. (“Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics” 2020)

  • Consider combining state and commonsense graph representations to achieve improved sample efficiency and results in text-based reinforcement learning tasks. (Adhikari et al. 2020)

  • Consider using a Breadth First Reasoning Graph (BFR-Graph) for multi-hop question answering tasks, as it presents a new message passing way that better conforms to the reasoning process and prevents unnecessary updates, leading to improved accuracy in answer span extraction. (Beltagy, Peters, and Cohan 2020)

  • Disaggregate data to uncover individual-level differences in behavior, which can provide valuable insights into the underlying mechanisms driving group-level phenomena. (Breuer, Eilat, and Weinsberg 2020)

  • Utilise the comprehensive knowledge graph completion benchmark, CoDEx, which offers improved scope and difficulty levels compared to existing benchmarks, providing a more accurate representation of real-world scenarios. (P. Jain, Rathi, and Chakrabarti 2020a)

  • Consider utilizing the AutoAudit framework for detecting anomalies in complex, time-evolving accounting datasets, as it offers scalability, generalizability, and improved interpretability through its Smurfing Detection, Attention Routing, Insight Discovery, and Scalability & Generalization components. (M.-C. Lee et al. 2020)

  • Utilise the Graph neural networks Including SparSe inTerpretability (GISST) model for interpreting important graph structure and node features in any GNN model, across various domains, due to its model-agnostic framework, attention mechanism, sparsity regularization, and end-to-end optimization. (C. Lin et al. 2020)

  • Consider using molecular counterfactuals to improve the interpretability of deep graph networks in molecule property prediction tasks, allowing domain experts to better understand the models decision-making processes.’ (Numeroso and Bacciu 2020)

  • Combine triple and path embedding and aggregation methods for fact checking, focusing on domain-relevant paths and utilizing deep learning models to assess the truthfulness of statements. (Pirrò 2020)

  • Leverage greedy search algorithms and zeroth-order methods to efficiently and effectively train robust Graph Neural Networks (GNNs) for large-scale problems. (Kaidi Xu et al. 2020)

  • Consider the coordinated nature of entity alignment tasks, where each alignment decision may highly correlate to the other decisions, and develop strategies like the Easy-to-Hard decoding strategy and joint entity alignment algorithm to improve overall performance. (Kun Xu et al. 2020)

  • Consider utilizing a combination of hierarchical attention and temporal attentive RNN models when dealing with dynamic heterogeneous networks, as they allow for the simultaneous capture of heterogeneous information and evolutionary patterns. (H. Xue et al. 2020)

  • Utilize a Temporal Event Graph Model for predicting event instances following a Temporal Complex Event Schema, which encapsulates events, arguments, temporal connections, and argument relations, leading to improved accuracy in event prediction tasks. (Yoo et al. 2020)

  • Not rely solely on structural information for entity alignment (EA), particularly for long-tail entities, and instead should incorporate additional signals such as entity names to improve accuracy. (Weixin Zeng et al. 2020)

  • Use the KEdge algorithm to improve the interpretability of graph neural networks (GNNs) by explicitly sparsifying the underlying graph, thereby reducing the impact of irrelevant neighbours on the prediction process. (“Machine Learning and Knowledge Discovery in Databases” 2020)

  • Consider implementing a hardware-based workload distribution autotuning framework, which includes an efficient online workload profiler and three workload rebalancing techniques, to effectively handle extreme workload imbalance in Graph Convolutional Networks (GCNs) and improve overall system efficiency. (Tianqi Wang et al. 2020)

  • Consider integrating heterogeneous modalities like traffic volume into your models through a domain transformer, while accounting for non-Euclidean spatial dependencies via graph convolution and a compound adjacency matrix that better reflects innate traffic proximity. (R. Dai et al. 2020)

  • Utilise the Graph Contrastive Coding (GCC) framework when attempting to learn structural graph representations across multiple graphs. (J. Qiu et al. 2020)

  • Consider using Adaptive Multi-Channel Graph Convolutional Networks (AM-GCN) when dealing with semi-supervised classification problems involving complex relationships between network data and classification tasks, as it allows for simultaneous learning of node embeddings based on node features, topological structures, and your combinations, while adaptively fusing the most relevant information for accurate classification. (Xiao Wang et al. 2020)

  • Utilise a combination of both instance-level and model-level methods to effectively explain deep graph neural networks, as these two types of methods offer complementary perspectives and insights. (Hao Yuan et al. 2020)

  • Consider incorporating contextual alignments when performing cross-lingual entity alignment tasks, as doing so allows for improved accuracy and reduced semantic gaps between different languages. (Z. Xie et al. 2020)

  • Consider using a heterogeneous graph neural network approach to better represent heterogeneous relations and capture discontinuous event segments in event chain analysis, leading to improved performance in one-step and multi-step inference tasks. (Jianming Zheng et al. 2020)

  • Consider using a combination of label-aware similarity measures, reinforcement learning-based neighbor selection, and relation-aware neighbor aggregators when working with Graph Neural Networks (GNNs) to improve fraud detection accuracy in cases involving camouflaging behaviors. (Y. Dou et al. 2020)

  • Consider developing a degree-specific GCN layer that uses an RNN-based parameter generator to effectively capture the local inter-relation of nodes with similar degrees, thereby reducing bias in GCNs caused by non-i.i.d node degrees. (X. Tang, Yao, et al. 2020)

  • Focus on developing a novel neural network-based model, Graph-aware Co-Attention Networks (GCAN), to accurately predict fake news on social media while providing explainability by highlighting suspicious retweeters and the specific words they focus on within the source tweet. (Yi-Ju Lu and Li 2020)

  • Consider combining Graph Convolutional Networks (GCNs) with Markov Random Field reasoning to improve social spammer detection, as demonstrated by your successful application on two real-world social network datasets. (Yongji Wu et al. 2020)

  • Use a collaboration-based multi-label propagation (CMLP) algorithm to effectively detect fraud users in e-commerce platforms, taking advantage of label correlations and improving scalability through a heterogeneous graph-based variant. (Haobo Wang et al. 2020)

  • Focus on developing randomized smoothing techniques specifically tailored for binary data structures, rather than relying on traditional Gaussian or Laplacian noise approaches, in order to effectively certify the robustness of community detection algorithms against adversarial structural perturbations. (J. Jia et al. 2020)

  • Consider maximising the mutual information between graph-level representations and the representations of substructures of different scales when developing unsupervised and semi-supervised graph-level representation learning methods. (“Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence” 2019)

  • Use the EdgeConv operation to effectively capture local geometric structure while maintaining permutation invariance in point cloud processing tasks such as classification and segmentation. (Yue Wang et al. 2019)

  • Utilise Graph Neural Networks (GNNs) as a computational substrate to model spatial processes without a predetermined graphical structure. By assigning nodes of a GNN to spatial locations and defining a computational process on the graph, researchers can model relationships between an initial function defined over a space and a resulting function in the same space. Furthermore, the locations of the nodes in space and your connectivity can be optimised to focus on the most complex parts of the space, (Alet et al. 2019)

  • Consider combining Graph Neural Networks (GNNs) and Conditional Random Fields (CRFs) in a mutual way to improve performance in few-shot learning tasks, as demonstrated through experimental results on miniImageNet and CIFAR-FS datasets. (K. R. Allen et al. 2019)

  • Focus on developing models that are equivariant to rotations, translations, reflections, and permutations, using the E(n)-Equivariant Graph Neural Networks (EGNNs) approach, which offers superior performance over traditional methods without requiring computationally expensive higher-order representations. (B. Anderson, Hy, and Kondor 2019)

  • Consider utilizing non-Euclidean embedding techniques, specifically those involving Riemannian manifolds, for analyzing complex temporal knowledge graphs due to your superior capability in capturing diverse geometric structures and temporal dynamics compared to traditional Euclidean approaches. (Bachmann, Bécigneul, and Ganea 2019)

  • Utilize a combination of pattern mining techniques and optimal transport on graphs to create comprehensible and justifiable interpretations of latent spaces in graph neural networks. (Baldassarre and Azizpour 2019)

  • Carefully consider the scale of your community detection attack, ranging from global to mesoscale to microscale, and develop tailored methods accordingly, such as the proposed Evolutionary Perturbation Attack (EPA) algorithm, which generates approximate optimal adversarial networks with minimal rewired links to launch all three scales of attacks. (Jinyin Chen et al. 2019)

  • Consider integrating logical operations such as count, superlative, aggregation, etc., into your fact-checking models to achieve improved accuracy and robustness. (Wenhu Chen, Wang, et al. 2019)

  • Adopt an “active learning” approach when dealing with heterogeneous networks, specifically through the development of a novel “Active Heterogeneous Network Embedding” (ActiveHNE) framework. This involves two main components: Discriminative Heterogeneous Network Embedding (DHNE) and Active Query in Heterogeneous Networks (AQHN). The former uses graph convolutional neural networks to embed the network, while the latter employs a multi-armed bandit (Xia Chen et al. 2019)

  • Consider using a novel scalable Deep Graph Bayesian Optimization (DGBO) method when working with attributed graphs, as it prevents the cubical complexity of Gaussian processes (GPs) by adopting a deep graph neural network to surrogate black-box functions, allowing it to scale linearly with the number of observations. (Jiaxu Cui, Yang, and Hu 2019)

  • Leverage PyTorch Geometric, a library specifically designed for deep learning on irregularly structured input data such as graphs, point clouds, and manifolds, to achieve high data throughput and efficiently handle input examples of different sizes. (Fey and Lenssen 2019)

  • Carefully consider the potential effects of structural noise on the performance of graph neural network models, and investigate ways to mitigate these effects through techniques like graph-augmented training. (J. Fox and Rajamanickam 2019)

  • Focus on the expressive power of aggregate-combine GNNs (AC-GNNs) and aggregate-combine-readout GNNs (ACR-GNNs) when analyzing graph neural networks, as these models can effectively capture various types of logical classifiers within the framework of first-order predicate logic. (Haonan et al. 2019)

  • Utilise an encoder-decoder framework for analysing dynamic graphs, which involves using an encoder to map nodes and relationships within the graph to hidden representations, and then applying a decoder to make predictions based on those representations. (Kazemi et al. 2019)

  • Consider using the Self-Attention Graph Pooling (SAGPool) method for graph pooling in Graph Neural Networks (GNNs) as it effectively combines node features and graph topology, leading to better performance in graph classification tasks. (Junhyun Lee, Lee, and Kang 2019)

  • Utilise a dynamic span graph framework when conducting information extraction tasks. This framework enables the propagation of global contextual information, improving the accuracy of results. (Y. Luan et al. 2019)

  • Utilize the Weisfeiler-Lehman test of isomorphism to develop scalable kernels for large graphs with discrete node labels, thereby improving the runtime and accuracy of graph kernels in various applications. (Togninalli et al. 2019)

  • Consider utilizing a deep generative model, specifically a variant of the structural constraint, to learn the DAG (Directed Acyclic Graph) in your studies. (Y. Yu et al. 2019)

  • Consider incorporating semantic hierarchies when developing knowledge graph embedding models, as doing so can lead to significant improvements in performance for link prediction tasks. (Liqiang Zhang et al. 2019)

  • Consider leveraging graph embedding techniques to integrate graph-structured data into a unified neural ranking framework, allowing for improved handling of query-item sparsity problems, generalization to unseen queries and long-tailed products, and fusion of external heterogeneous information to enhance search results. (Yuan Zhang, Wang, and Zhang 2019)

  • Develop a comprehensive graph neural network platform like AliGraph, which offers distributed graph storage, optimized sampling operators, and runtime to efficiently support existing and newly developed GNN algorithms across diverse scenarios. (R. Zhu et al. 2019)

  • Focus on developing a comprehensive understanding of the various factors contributing to fraudulence in complex networks, rather than solely relying on traditional density-based methods. (Shenghua Liu, Hooi, and Faloutsos 2019)

  • Utilise the SubgraphX method to effectively explain the predictions of graph neural networks (GNNs) by identifying important subgraphs through efficient exploration using Monte Carlo tree search and measuring subgraph importance with Shapley values. (Devlin et al. 2019)

  • Utilize advanced computational methods like those borrowed from condensed matter physics to effectively study spectral densities in networks, allowing for scalability and efficiency even in extremely large graphs. (K. Dong, Benson, and Bindel 2019)

  • Focus on developing end-to-end solutions for the exact-K recommendation problem, reducing it to a Maximal Clique Optimization problem, and leveraging Graph Attention Networks (GAtN) with an Encoder-Decoder framework to effectively address this challenging issue. (Yu Gong et al. 2019)

  • Consider using Multi-LENS, an inductive multi-level latent network summarization approach, to address the challenge of large and dense graph structures by creating a compact, size-independent representation of the graph structure through the use of relational operators and functions, leading to improved performance in applications like link prediction and event detection. (D. Jin et al. 2019)

  • Develop a framework called “Multiple Conditional Network Embeddings” (MCNE) to learn multiple conditional node representations to represent multiple aspects of similarity between nodes within a single vector space. (Hao Wang et al. 2019)

  • Consider incorporating degree-specific graph-level pooling methods in your studies to improve the accuracy and robustness of your results. (Jun Wu, He, and Xu 2019)

  • Utilise Gaussian distributions in graph convolutional layers to absorb the effects of adversarial attacks and introduce a variance-based attention mechanism to prevent the propagation of adversarial attacks in GCNs. (D. Zhu et al. 2019)

  • Focus on developing a probabilistic model, like the Dual-Task Factor Graph (DTF), to jointly identify default borrowers and cheating agents in a given mobile network using various factors such as user features, cheating agent features, and the correlation between default borrower identity and cheating agent identity. (Yang Yang et al. 2019)

  • Utilise an Attribute Heterogeneous Information Network (AHIN) to model complex systems involving multiple types of entities and relationships, and then employ a combination of Graph Convolutional Networks (GCN) and attention mechanisms to learn effective embeddings for downstream tasks like node classification. (Yiming Zhang, Fan, Ye, et al. 2019)

  • Consider utilizing a hierarchical attention mechanism within an Attributed Heterogeneous Information Network (AHIN) when dealing with cash-out user detection problems. (B. Hu et al. 2019)

  • Utilise a combination of static neighbour encoders and graph neural network-based recurrent units to effectively capture both temporal and static interaction patterns in node classification tasks. (H. Park and Neville 2019)

  • Utilise a combination of static node embeddings and temporal node embeddings to accurately predict future interactions within a temporal graph. (U. Singer, Guy, and Radinsky 2019)

  • Develop a vectorized relational graph convolutional network (VR-GCN) to learn the embeddings of both graph entities and relations simultaneously for multi-relational networks, enabling improved network embedding, entity alignment, and relation alignment. (R. Ye et al. 2019)

  • Incorporate neighborhood subgraph-level information of entities in addition to triplet-based knowledge when performing entity alignment across multilingual knowledge graphs, as this leads to improved performance compared to state-of-the-art entity alignment models. (Qiannan Zhu et al. 2019)

  • Employ Deep Divergence Graph Kernels (DDGK) to effectively learn unsupervised representations of graphs and your nodes without depending on feature engineering or domain knowledge, while leveraging an innovative isomorphism attention mechanism for cross-graph attention to enhance interpretability and discovery of similar substructures. (Al-Rfou, Perozzi, and Zelle 2019)

  • Adopt a novel graph neural network framework (GraphRec) for social recommendations, which addresses the challenges of integrating multiple graphs, capturing interactions and opinions, and distinguishing social relations with varying strengths. (W. Fan et al. 2019)

  • Consider utilizing graph convolutional networks (GCNs) and graph pooling (gPool) layers in conjunction with hybrid convolutional (hConv) layers for text modeling tasks, as demonstrated by the superior performance of the hConv-gPool-Net network compared to alternative approaches across multiple datasets. (H. Gao, Chen, and Ji 2019)

  • Consider using Knowledge Graph Convolutional Networks (KGCN) for recommender systems, as they effectively capture high-order structure and semantic information within the knowledge graph, leading to improved performance compared to other state-of-the-art models. (Hongwei Wang, Zhao, et al. 2019)

  • Integrate both writing and photography styles when developing an intelligent system for automatic linking of multiple accounts belonging to the same individuals involved in drug trafficking in darknet markets. (Yiming Zhang, Fan, Song, et al. 2019)

  • Consider using TuckER, a linear model based on Tucker decomposition, for knowledge graph completion tasks, as it outperforms previous state-of-the-art models, is fully expressive, and serves as a strong baseline for more elaborate models. (Balazevic, Allen, and Hospedales 2019)

  • Leverage complex network properties to improve the robustness of graph convolutional networks (GCNs) in the presence of adversaries, focusing on training data selection strategies that prioritize well-connected nodes. (Kegelmeyer, Wendt, and Pinar 2018)

  • Consider utilising graph representations and extracting graph features for developing effective and precise time series classification (TSC) algorithms. (D. Li et al. 2018)

  • Consider adopting a junction tree encoder-decoder framework for learning diverse graph translations, combined with a novel adversarial training method for aligning distributions of molecules, to effectively optimize molecular properties. (Gómez-Bombarelli et al. 2018)

  • Consider reducing the complexity of Graph Convolutional Networks (GCNs) by removing nonlinearities and collapsing weight matrices between consecutive layers, leading to a linear model that corresponds to a fixed low-pass filter followed by a linear classifier. This approach maintains accuracy in many downstream applications, scales better to larger datasets, provides greater interpretability, and offers significant speedups compared to existing models like FastGCN. (Abu-El-Haija et al. 2018)

  • Utilize a hybrid approach combining traditional sequence encoders with graph neural networks to improve the performance of automatic summarization tasks across various domains. (Allamanis 2018)

  • Develop novel semi-supervised network embedding methods specifically tailored to handle completely-imbalanced labels, where some classes have no labeled nodes at all, rather than relying solely on traditional semi-supervised methods that assume balanced labels. (Battaglia et al. 2018)

  • Focus on learning in the space of algorithms rather than just directly recovering solutions from raw inputs, as this enables better understanding of the underlying processes and facilitates positive transfer between tasks. (Battaglia et al. 2018)

  • Consider using MeshGraphNets, a framework for learning mesh-based simulations using graph neural networks, to accurately predict the dynamics of various physical systems, including aerodynamics, structural mechanics, and cloth, while improving efficiency through adaptability and scalability. (Battaglia et al. 2018)

  • Consider using Deep Graph Infomax (DGI) for unsupervised learning of node representations within graph-structured data, as it relies on maximizing mutual information between patch representations and corresponding high-level summaries of graphs, leading to improved performance compared to existing methods. (Battaglia et al. 2018)

  • Consider using Gated Graph Neural Networks (GGNNs) coupled with an input transformation that enables nodes and edges to have separate hidden representations, thereby allowing for improved encoding of graph structures in various NLP applications. (D. Beck, Haffari, and Cohn 2018)

  • Consider using a learning-based approach to detect and fix a broad range of bugs in JavaScript programs by framing the problem in terms of learning a sequence of graph transformations, rather than relying solely on deep neural networks or rule-based methods. (Brockschmidt et al. 2018)

  • Develop a comprehensive neural network model capable of capturing multiple properties of traffic data, including region-level correlations, temporal periodicity, and inter-traffic correlations, to improve the accuracy of joint predictions for travel demands and traffic flows across all regions of a city. (Ken Chen et al. 2018)

  • Consider incorporating edge labels in addition to node labels when developing network embedding models, as doing so allows for more accurate predictions of node attributes. (H. Chen et al. 2018)

  • Leverage the social context through which information is propagated to enhance the representation of the information and provide a distant-supervision source based on users who endorse and spread the content. (Glenski, Weninger, and Volkova 2018)

  • Ensure proper separation of model selection and model assessment phases, avoiding over-optimistic and biased estimates of model performance, and promoting transparency and reproducibility in experimental settings. (Lipton and Steinhardt 2018)

  • Consider implementing GraphIE, a graph-based framework that uses graph convolutions to propagate information between connected nodes, resulting in improved word-level predictions compared to traditional sequential tagging models. (Y. Qian et al. 2018)

  • Consider utilizing Graph Neural Networks (GNNs) for solving inference problems in probabilistic graphical models, particularly in situations involving loopy graphs, as GNNs have been shown to significantly outperform traditional belief propagation techniques in these scenarios. (K. Yoon et al. 2018)

  • Adopt a Bayesian approach to view the observed graph as a realization from a parametric family of random graphs, and then target inference of the joint posterior of the random graph parameters and the node (or graph) labels. (Yingxue Zhang et al. 2018)

  • Consider the importance of preserving both local and global network structure, effectively utilizing vertex attributes, addressing data sparsity issues, and ensuring scalability when developing network representation learning algorithms. (Daokun Zhang et al. 2018)

  • Treat spectral graph wavelets as probability distributions and characterize your distributions using empirical characteristic functions, enabling accurate recovery of structurally similar and structurally equivalent nodes in graphs. (Donnat et al. 2018)

  • Utilize the HEER algorithm to address the challenge of comprehensive transcription of heterogeneous information networks (HINs) by incorporating edge representations and heterogeneous metrics, thus enabling accurate and efficient learning from networked data. (Yu Shi et al. 2018)

  • Carefully consider the impact of adversarial attacks on neural networks for graph data, particularly in areas where such attacks may be common, and explore strategies for generating adversarial perturbations that remain undetectable while effectively altering the graph structure and node features. (Zügner, Akbarnejad, and Günnemann 2018)

  • Focus on developing and testing embedding-based knowledge graph completion methods on more challenging, realistic datasets like FB15k-237, rather than relying solely on simpler datasets like FB15k, to ensure robustness and generalizability of your models. (Akrami et al. 2018)

  • Integrate both entity attributes and second-order structures within a deep heterogeneous network embedding approach to effectively identify abnormal events in Heterogeneous Information Networks (HINs). (S. Fan, Shi, and Wang 2018)

  • Consider implementing L2 regularization on both input and output vectors in order to address the norm convergence problem in SGNS-based network embedding algorithms, thereby improving the quality of the resulting embeddings. (Yi Zhang, Lu, and Shai 2018)

  • Consider extending Skip-gram based network embedding methods to dynamic settings, allowing for efficient updating of vertex representations as networks evolve. (L. Du et al. 2018)

  • Consider applying the k-core decomposition technique when analyzing graph similarity across various scales, as it enables a more comprehensive understanding of the graphs structure and improves classification accuracy.’ (Nikolentzos et al. 2018)

  • Consider using self-attention graph pooling (SAGPool) for graph neural network (GNN) applications, as it enables efficient hierarchical representation learning while considering both node features and graph topology. (Rhee, Seo, and Kim 2018)

  • Use the MASTER framework, which integrates attribute and structural embedding across multiple social networks, to overcome limitations such as multiplicity, comprehensiveness, and robustness when reconciling social networks. (S. Su et al. 2018)

  • Utilise a scalable multiplex network embedding model to effectively represent information from multi-type relations into a unified embedding space. (Hongming Zhang et al. 2018)

  • Utilize large-scale commonsense knowledge in your conversation generation models, specifically by employing static and dynamic graph attention mechanisms to effectively encode and decode the semantic information within the knowledge graphs. (Hao Zhou et al. 2018)

  • Leverage a dynamic index structure to optimize trade-offs between memory usage and query efficiency, with minimized maintenance cost, specifically proposing a hot point based index’, which can be selectively applied to a certain portion of the graph and exploit various heuristics (such as memory usage and vertex connectivities) to balance cost and efficiency. (Xiafei Qiu et al. 2018)

  • Carefully choose a distance metric when comparing graphs, considering whether they want to focus on local, mesoscale, or global structural changes, as different metrics may perform differently depending on the desired scale. (Donnat and Holmes 2018)

  • Consider casting few-shot learning as a supervised message passing task on a graph, which can be effectively implemented using graph neural networks. (Garcia and Bruna 2017)

  • Utilize graph signal processing (GSP) to effectively analyze and interpret complex data sets that exist on irregular graph domains, extending classical signal processing concepts like Fourier transform, filtering, and frequency response to better understand these unique data structures. (Ortega et al. 2017)

  • Consider incorporating nodewise predictability measures in your network models to better understand the practical relevance of edges and improve the effectiveness of potential interventions. (Haslbeck and Waldorp 2017)

  • Utilize a standardized, open-source GNN benchmarking framework to ensure fair and consistent comparisons between different GNN models, allowing for easier identification of effective architectures and techniques. (Monti et al. 2017)

  • Utilize a combination of adversarial filters and compositional strategies to create flexible, invariant embeddings that can effectively handle various combinations of sensitive attributes in graph representation learning. (Berg, Kipf, and Welling 2017)

  • Consider using Graph2Gauss, a novel approach that embeds nodes as Gaussian distributions, enabling the capture of uncertainty in node representations, and employs an unsupervised personalized ranking formulation to effectively capture the non-i.i.d. nature of the data arising from complex interactions between nodes. (Bojchevski and Günnemann 2017)

  • Incorporate the degree penalty’ principle in your network embedding algorithms to effectively preserve the scale-free property of networks, thereby improving the quality of vertex representations.’ (R. Feng et al. 2017)

  • Carefully study and optimize the use of Message Passing Neural Networks (MPNNs) for supervised learning on molecular graphs, focusing on improving your performance on relevant benchmarks and understanding your limitations, in order to establish them as the go-to solution for chemical prediction problems. (Gilmer et al. 2017)

  • Consider using hyperbolic spaces instead of Euclidean spaces for heterogeneous information network (HIN) embedding, as hyperbolic spaces can better capture the hierarchical and power-law structure present in HINs. (Zhipeng Huang and Mamoulis 2017)

  • Adopt a semi-supervised approach when dealing with entity alignment tasks, combining both knowledge embedding models and cross-graph models to effectively handle incomplete knowledge graphs and improve the accuracy of entity alignment. (Kotnis and Nastase 2017)

  • Consider using graph-to-sequence neural networks (GraphMR) for mathematical reasoning tasks, as it outperforms traditional sequence-to-sequence models in handling structural information and improves overall performance. (Ling et al. 2017)

  • Consider using dynamic edge-conditioned filters in convolutional neural networks on graphs to improve graph classification performance, particularly in situations where edge labels provide valuable information about the relationships between nodes. (Simonovsky and Komodakis 2017)

  • Consider integrating an initial scores learning algorithm for non-seed nodes with side information into a graph-based propagation method to significantly improve the accuracy of propagation on large systems where precise labels are rare. (Jinlong Hu, Liang, and Dong 2017)

  • Focus on extending your analysis beyond just boundary points to include interior points within the domain, as doing so provides a more comprehensive understanding of the imaginary geometry being investigated. (J. Miller and Sheffield 2017)

  • Consider combining different types of proximity measures (first-, second-, and high-order) in your node embedding techniques to improve community detection and embedding performance. (Cavallari et al. 2017)

  • Utilise a combination of offline and online models to efficiently analyse dynamic attributed networks, whereby an initial offline model is established to create a consensus embedding, followed by an online model that updates this embedding using matrix perturbation theory to account for changes in network structure and attributes. (Jundong Li et al. 2017)

  • Consider incorporating both signed links and user attributes when analyzing signed social networks, as doing so can lead to improved network representation and better performance in tasks like link prediction and node clustering. (Suhang Wang et al. 2017)

  • Carefully consider the limitations of existing node representation learning algorithms like DeepWalk and node2vec, particularly in regards to your ability to accurately capture structural equivalence, and explore alternatives like struc2vec that prioritize structural identity. (L. F. R. Ribeiro, Saverese, and Figueiredo 2017)

  • Develop a task-guided and path-augmented heterogeneous network embedding model to improve the accuracy of author identification in double-blind review settings. (Ting Chen and Sun 2017)

  • Consider incorporating graph-based methods, semi-supervised learning, and domain-specific knowledge to improve the performance of fraud detection systems in e-commerce applications. (“Complex Networks &Amp; Their Applications v” 2017)

  • Utilise the concept of Frechet means to define the average of network data objects, enabling them to conduct one- and two-sample tests for network data. (Ginestet et al. 2017)

  • Focus on developing graph neural network (GNN) architectures that can effectively capture distinctive and stable local structures within complex graphs, while being scalable and computationally efficient. (Abbe 2017)

  • Utilize the Gromov-Wasserstein learning framework to jointly optimize graph matching and node embedding tasks, thereby enhancing the accuracy and efficiency of both processes. (Altschuler, Weed, and Rigollet 2017)

  • Consider utilizing graph-based generative procedures when dealing with highly structured objects, as it allows for the incorporation of rich structural information. (Amodio, Chaudhuri, and Reps 2017)

  • Leverage implicit and explicit relational knowledge from knowledge graph embeddings and graph convolution networks to effectively deal with long-tailed, imbalanced data in relation extraction tasks. (Bastings et al. 2017)

  • Adopt the novel GNN Self-Distillation (GNN-SD) technique for training Graph Neural Networks (GNNs) because it enables effective knowledge transfer within a single GNN model without requiring external teacher models, thereby reducing training costs and enhancing overall performance. (Bresson and Laurent 2017)

  • Carefully consider the specific problem setting and corresponding challenges when developing graph embedding techniques, as different types of inputs and outputs lead to distinct difficulties and opportunities for improvement. (Hongyun Cai, Zheng, and Chang 2017)

  • Utilize a broad learning approach called HitFraud’, which uses heterogeneous information networks (HINs) to detect fraudulent behavior in payment transactions. This approach improves upon traditional methods by identifying inter-transaction dependencies, thus allowing for the detection of correlated and rapidly changing fraudulent activities.’ (B. Cao et al. 2017)

  • Employ a hierarchical representation learning framework like HARP to enhance the quality of graph embeddings, thereby improving downstream task performances such as classification. (H. Chen et al. 2017)

  • Consider implementing a control variate based algorithm to effectively reduce the variance of your estimators, leading to more accurate predictions and faster convergence rates. (Jianfei Chen, Zhu, and Song 2017)

  • Consider using GraphSAGE, a general inductive framework that leverages node feature information to efficiently generate node embeddings for previously unseen data, allowing for better generalization to unseen nodes and improved performance in various prediction tasks. (Jianfei Chen, Zhu, and Song 2017)

  • Carefully consider the impact of meta-path selection on the quality of learned node embeddings in heterogeneous graph embeddings, and explore alternative solutions like the proposed JUST method that uses random walks with jump and stay strategies instead of meta-paths. (P. Cui et al. 2017)

  • Utilise a collaborative policy learning (CPL) framework to jointly train two reinforcement learning (RL) agents - a multi-hop graph reasoner and a fact extractor - in order to enhance open knowledge graph reasoning tasks. (R. Das et al. 2017)

  • Consider integrating multiple feature types, including latent, relational, and numerical features, into your knowledge base representations through the use of a novel combination of neural representation learning and probabilistic product of experts models. (Dettmers et al. 2017)

  • Consider employing a post-processing rule-based companion to sub-symbolic explainer methods in order to aggregate global rule-based explanations through a standard white-box machine learning technique, thereby reducing the amount of interpretation required by users and providing a model-level explanation that captures the global behaviour of a model. (Doran, Schulz, and Besold 2017)

  • Consider using the disentangled graph convolutional network (DisenGCN) to address challenges in learning disentangled node representations, as it effectively utilizes a novel neighborhood routing mechanism to identify the factor causing links between nodes and enables accurate extraction of features specific to those factors. (Doshi-Velez and Kim 2017)

  • Consider casting few-shot learning as a supervised message passing task which is trained end-to-end using graph neural networks. (Garcia and Bruna 2017)

  • Utilise the proposed graph pooling (gPool) and unpooling (gUnpool) operations to effectively handle graph data in encoder-decoder architectures, leading to improved performance in node classification and graph classification tasks. (Gilmer et al. 2017)

  • Consider using GraphSAGE, a general inductive framework that leverages node feature information to efficiently generate node embeddings for previously unseen data, allowing for better generalization to unseen nodes and improved performance in various prediction tasks. (Hamilton, Ying, and Leskovec 2017a)

  • Exploit node content for multiview graph convolutional networks and utilize adversarial regularization to improve the accuracy and reliability of your network analysis. (Hamilton, Ying, and Leskovec 2017b)

  • Adopt an encoder-decoder framework when developing node embedding methods for representation learning on graphs, which involves defining a pairwise similarity function, an encoder function, a decoder function, and a loss function to optimize the encoder-decoder system. (Hamilton, Ying, and Leskovec 2017b)

  • Consider using VAIN, a novel attentional architecture for multi-agent predictive modeling, because it scales linearly with the number of agents, effectively models high-order interactions, and outperforms competing multi-agent approaches in various domains like chess and soccer. (Hoshen 2017)

  • Consider using Graph Neural Networks (GNNs) to model dependencies between roles and improve performance in situation recognition tasks, as demonstrated by the significant improvements achieved in the study. (Ruiyu Li et al. 2017)

  • Utilise a novel framework called “Semi-supervised Embedding in Attributed Networks with Outliers” (SEANO) to learn a low-dimensional vector representation that systematically captures the topological proximity, attribute affinity, and label similarity of vertices in a partially labeled attributed network (PLAN). (J. Liang et al. 2017)

  • Consider leveraging analogical inference in your studies, specifically through the use of multi-relational embedding techniques, to better understand and predict relationships between entities and your interactions. (Hanxiao Liu, Wu, and Yang 2017)

  • Consider adopting a probabilistic framework for knowledge graph embeddings, allowing them to interpret these models in a way that facilitates efficient hyperparameter tuning through a variational expectation-maximization approach. (D. Q. Nguyen 2017)

  • Utilize the proposed structure2vec’ method for efficient and accurate handling of structured data, particularly in scenarios involving millions of data points, due to its ability to run twice as fast, produce models 10,000 times smaller, and maintain state-of-the-art predictive performance.’ (H. Dai, Dai, and Song 2016)

  • Utilise the Kronecker graph model for generating realistic networks due to its ability to capture multiple properties of real networks simultaneously, while remaining analytically tractable and amenable to rigorous analysis. (Hamilton et al. 2016)

  • Consider developing and evaluating supervised algorithms for embedding graphs in hyperbolic space, as well as combining Euclidean and hyperbolic embeddings for improved representational power in node classification and link prediction tasks. (Kipf and Welling 2016a)

  • Utilise interaction networks, a model that separates reasoning about relations from reasoning about objects, assigns each task to distinct models, and allows automatic generalisation across variable numbers of arbitrarily ordered objects and relations. (Battaglia et al. 2016)

  • Consider integrating feature diffusion and graph node embedding simultaneously into a unified network through a novel diffusion-embedding architecture when working with graph-structured data. (Kipf and Welling 2016a)

  • Consider developing a novel heterogeneous graph neural network (HmGNN) to effectively detect fraudulent invitations in online internet enterprises, while simultaneously addressing the challenges of large yet locally small graph structures and heterogeneous user associations. (Kipf and Welling 2016b)

  • Utilize a graph-based method for anomaly detection in public procurement, specifically the PANG (Pattern-Based Anomaly Detection in Graphs) framework, which enables the identification of induced subgraphs and improves overall predictive performance. (Acosta-Mendoza et al. 2016)

  • Incorporate the concept of “triadic closure” when studying dynamic networks, as it allows for better understanding of network evolution and individual behavior within the network. (Linhong Zhu et al. 2016)

  • Develop a class of density metrics for bipartite graphs that can be optimized in near-linear time, within a constant factor of the optimum, and is minimally affected by camouflage edges added by adversaries, in order to effectively detect fraud in social networks. (Hooi et al. 2016)

  • Focus on developing a structural neighborhood-based classifier learning using a random walk, emphasizing the role of short random walks for effective classification in sparse and noisy networks. (Nandanwar and Murty 2016)

  • Consider utilizing graph neural networks when dealing with structured data, as it allows for efficient processing and improved accuracy. (Andreas, Klein, and Levine 2016)

  • Consider using the Relational-variational Graph Autoencoder (R-VGAE) model for unsupervised prerequisite chain learning, as it outperforms graph-based semi-supervised methods and other baseline methods by up to 9.77% and 10.47% in terms of prerequisite relation prediction accuracy and F1 score. (Shraey Bhatia, Lau, and Baldwin 2016)

  • Consider using the “joint graph decomposition and node labeling” approach when dealing with complex computer vision tasks like multiple object tracking, instance-separating semantic segmentation, and articulated human body pose estimation, as it provides a common mathematical abstraction and efficient algorithms for solving these problems. (L.-C. Chen et al. 2016)

  • Consider combining global structure and local semantics when performing entity alignment tasks, as doing so can lead to improved accuracy and robustness in identifying entities across different knowledge graphs. (Muhao Chen et al. 2016)

  • Utilize graph query embeddings (GQEs) to efficiently make predictions about conjunctive queries on incomplete knowledge graphs. (W. W. Cohen 2016)

  • Formulate the combinatorial problem of graph matching as an Integer Linear Programming (ILP) problem, which allows for greater flexibility and efficiency when comparing graphs of varying sizes. (Kipf and Welling 2016a)

  • Utilise the Table2Graph framework when dealing with tabular data, as it enables the transformation of feature interaction modeling into a unified graph learning problem. This approach addresses the challenge of effectively learning a unified feature-interaction graph by using reinforcement learning to stabilise key feature interaction connections and proposing a differentiable sparsity constraint to regulate edge connections. The result is improved prediction accuracy and feature interaction detection, making the process more efficient than alternative methods. (Kipf and Welling 2016a)

  • Utilise a novel task-agnostic explanation pipeline for graph neural networks (GNNs) that decomposes a prediction model into a GNN embedding model and a downstream model, enabling the explanation of multiple downstream tasks with a single embedding explainer. (Kipf and Welling 2016b)

  • Utilize a global recursive neural parsing model with optimality guarantees during decoding, which can be achieved by giving up dynamic programs and searching directly in the space of all possible subtrees, even though this space is exponentially large in the sentence length. (Kenton Lee, Lewis, and Zettlemoyer 2016)

  • Consider using the minimum cost node labeling lifted multicut problem (nl-lmp) as a unifying mathematical abstraction for various computer vision tasks, such as multiple object tracking, instance-separating semantic segmentation, and articulated human body pose estimation. (Levinkov et al. 2016)

  • Consider integrating structured prior knowledge in the form of knowledge graphs into your image classification pipelines, as this can lead to improved performance. (Marino, Salakhutdinov, and Gupta 2016)

  • Use a translation-based network representation learning approach, like TransNet, to effectively model and predict social relations in social networks, as it significantly outperforms traditional network representation learning methods and knowledge graph embedding approaches. (C. Tu et al. 2016)

  • Utilise persistence images’, a finite-dimensional vector representation of a persistence diagram, as it enables the application of a wide array of machine learning techniques, is stable with respect to input noise, computationally efficient, maintains an interpretable link to the original persistence diagram, and allows for adjustment of the relative importance of points in different regions of the persistence diagram.’ (Y.-C. Chen et al. 2015)

  • Carefully consider the role of graph connections in ensuring fairness in your analysis, particularly in relation to dyadic fairness, which requires independence of sensitive attributes in predictive relationships. (Edwards and Storkey 2015)

  • Develop a two-stream adaptive graph convolutional network (2s-AGCN) for skeleton-based action recognition, allowing for simultaneous modelling of first-order and second-order information, thereby improving recognition accuracy. (Henaff, Bruna, and LeCun 2015)

  • Consider adapting existing explainability methods for convolutional neural networks (CNNs) to graph convolutional neural networks (GCNNs) in order to better understand and interpret the predictions made by these models. (Henaff, Bruna, and LeCun 2015)

  • Integrate a learning component into the process of enriching knowledge graphs with external text, allowing for improved quality of final products, i.e., low-dimensional embeddings, by introducing new features obtained from a distinct knowledge source and established based on affinity captured by the learning component of the model. (G. Hinton, Vinyals, and Dean 2015)

  • Use the G-XAIBench library to systematically evaluate and compare the quality of various GNN explainers on both synthetic and real-world graphs using different performance metrics to quantify the quality of explanations. (Jordan and Freiburger 2015)

  • Focus on developing models that can effectively capture and utilize the rich structural information present in graph attention networks, thereby enabling accurate inferences about structural interactions between nodes and ultimately improving overall performance. (T. Luong, Pham, and Manning 2015)

  • Utilise the public-private graph model to develop efficient algorithms for solving complex graph problems in large-scale social networks, taking into account both the public and private connections among individuals. (Chierichetti et al. 2015)

  • Focus on developing algorithms that can efficiently solve the steady-state inversion problem, which involves finding the unique solution to a system of linear equations representing the steady-state distribution of a Markov chain, using a continuous, monotonic, and unbounded mapping from node scores to transition probabilities. (R. Kumar et al. 2015)

  • Consider adopting a human-in-the-loop approach for generating concept-based explanations for graph neural networks, as it enables better understanding and trust in the model predictions. (Abu-Aisheh et al. 2015)

  • Consider utilizing graph attention networks (GAT-CADNet) for solving panoptic symbol spotting problems in CAD drawings, as it effectively combines semantic and instance symbol spotting tasks in one consolidated network. (A. X. Chang et al. 2015)

  • Consider utilizing a closed-form solution for belief propagation, which offers improved scalability and effectiveness over traditional approaches, particularly in large-scale graph problems. (Gatterbauer 2015)

  • Consider implementing the DropEdge technique, which involves randomly removing a certain percentage of edges from the input graph at each training epoch, to improve the performance of deep Graph Convolutional Networks (GCNs) for node classification by addressing over-fitting and over-smoothing issues. (Henaff, Bruna, and LeCun 2015)

  • Focus on developing deep attention diffusion graph neural networks (DADGNN) for text classification, which addresses the limitations of conventional graph-based models by leveraging attention diffusion techniques to capture long-range word interactions, decoupling propagation and transformation processes for deeper networks, and calculating node weights for precise document-level representations. (Zhiheng Huang, Xu, and Yu 2015)

  • Use MixHop, a novel graph convolutional architecture that enables higher-order message passing and neighborhood mixing, allowing for more accurate and flexible representation of complex graph structures. (Ioffe and Szegedy 2015)

  • Consider jointly learning entity and relation representations for entity alignment, as doing so can significantly enhance the accuracy and efficiency of the alignment process. (Srivastava, Greff, and Schmidhuber 2015)

  • Ensure that your experimental designs account for the possibility of confounding factors and implement appropriate controls to minimize your impact on the study outcomes. (Bourgade, Erdös, and Yau 2014)

  • Focus on understanding the relationship between the planted partition model and the Erdős-Rényi model, particularly in terms of mutual contiguity and asymptotic orthogonality, in order to develop effective algorithms for clustering and parameter estimation in sparse graphs. (Mossel, Neeman, and Sly 2014)

  • Develop an online algorithm for recovering the weights over the edges that determine the graph in order to gain full understanding and interpretability of the social networks decision-making process.’ (Sayed 2014)

  • Utilise Relational Pooling (RP) as a novel framework for graph representation that can be combined with any existing neural network architecture, including those not typically associated with graphs such as Recurrent Neural Networks (RNNs). (K. Cho, Merrienboer, Gulcehre, et al. 2014)

  • Consider leveraging all three types of information - data structure, domain label, and class label - together within a unified deep learning model for optimal unsupervised domain adaptation. (Ajakan et al. 2014)

  • Consider incorporating geometric structure in the form of graphs into Hawkes processes to improve model prediction accuracy and efficiency. (J. Chung et al. 2014)

  • Utilise graph wavelet neural networks (GWNN) rather than spectral graph convolutional neural networks (CNNs) because GWNNs provide higher efficiency, sparseness, and localisation in vertex domains compared to spectral CNNs. (Diederik P. Kingma and Ba 2014)

  • Utilize a combination of semantic information from word embeddings and syntactic information from dependency graphs when attempting to accurately classify comparative preferences in sentences. (Diederik P. Kingma and Ba 2014)

  • Consider using core-guided MaxSAT algorithms instead of traditional iterative ones because they generally perform better and are more competitive against other methods. (Morgado et al. 2013)

  • Consider developing a novel data-free knowledge distillation approach specifically tailored for graph neural networks (GNNs), rather than attempting to apply existing methods developed for convolutional neural networks (CNNs) to GNNs. (Bruna et al. 2013)

  • Re-formalize the task of aspect-category based sentiment analysis as a category-sentiment hierarchy prediction problem, using a two-layer hierarchy output structure to improve the accuracy of aspect category detection and category-oriented sentiment classification. (Bruna et al. 2013)

  • Consider using AnoMulY, a general, unsupervised edge anomaly detection framework specifically designed for multiplex dynamic networks, which leverages node embeddings at different GNN layers as hierarchical node states and employs a GRU cell to capture temporal properties of the network and update node embeddings over time, while adding an attention mechanism to incorporate information across different types of relations. (Akoglu and Faloutsos 2013)

  • Consider utilizing a spectral graph theoretical formulation of convolutional neural networks (CNNs) on graphs, which allows for strict localization of filters within a specified radius, low computational complexity, and efficient pooling strategies. (Bruna et al. 2013)

  • Utilize a dual-discriminative graph neural network approach for imbalanced graph-level anomaly detection tasks, combining both anomalous attribute-aware graph convolution and anomalous substructure-aware deep Random Walk Kernel (deep RWK) techniques, along with a Point Mutual Information (PMI)-based loss function to handle the imbalance distribution issue. (Bruna et al. 2013)

  • Consider implementing a self-supervised semantic alignment graph convolution network (SelfSAGCN) to overcome limitations in graph convolution networks (GCNs) related to insufficient labeled data and indistinguishable features caused by multiple layers. This involves combining identity aggregation and semantic alignment techniques to map node features from both semantic and graph structural aspects, thereby improving classification performance. (Bruna et al. 2013)

  • Consider using a multi-layer multiplex graph neural net architecture for abstract diagram reasoning, as it enables the encoding of subsets of diagram panels into multi-layer multiplex graphs and the combination of summaries of several graphs to predict the correct candidate answer. (Eigen, Ranzato, and Sutskever 2013)

  • Consider incorporating a document relationship graph into your neural topic modelling approaches, as doing so allows for richer document and word representations and improved topic inference. (Diederik P. Kingma and Welling 2013)

  • Consider using a novel community structure embedding method to better capture inherent community structures in attributed graphs, rather than solely relying on traditional network topology measures. (Mikolov, Chen, et al. 2013)

  • Pay careful attention to the order-dependence of the PC-algorithm, especially when working with high-dimensional data, as it can lead to significant variation in results and conclusions across different variable orderings. (Colombo and Maathuis 2012)

  • Carefully consider the locality principle when studying self-interacting systems, as it provides valuable insights into the behavior of complex systems. (Dumaz and Tóth 2012)

  • Use a generative probabilistic model to infer the connectivity and transmission rates of diffusion networks, while considering multiple parametric models for transmission likelihoods, such as exponential, power-law, and Rayleigh, to capture diverse patterns in real-world data. (Rodriguez, Balduzzi, and Schölkopf 2011)

  • Utilise the random-cluster model on a finite connected graph, which is a model on the edges of the graph, each one being either closed or open, with the probability of a configuration being proportional to the edge-weight, cluster-weight, and number of clusters. (Beffara and Duminil-Copin 2011)

  • Ensure that the local semicircle law is established before attempting to apply the Greens function comparison theorem, as this allows for the removal of continuity and LSI restrictions and ultimately leads to the bulk universality for generalized Wigner matrices.’ (Erdős, Yau, and Yin 2011)

  • Develop a framework for learning convolutional neural networks for arbitrary graphs, enabling efficient and accurate processing of graph data for various tasks. (Douglas 2011)

  • Utilise the proposed generalisation of the Marcenko-Pastur equation to better understand the relationship between sample and population covariance matrices, enabling them to develop improved estimators of the covariance matrix and its inverse. (Ledoit and Péché 2010)

  • Utilize a temporally smoothed \(l_{1}\)-regularized logistic regression framework to effectively estimate time-varying networks from time series of entity attributes. (Kolar et al. 2010)

  • Consider utilizing local graph clustering algorithms for detecting fraudulent accounts in large datasets, as demonstrated by the success of the GraphRAD system in identifying previously unknown fraud accounts. (Akoglu, McGlohon, and Faloutsos 2010)

  • Carefully choose between Relational Neural Networks (RelNNs) and Graph Neural Networks (GNNs) when working with relational data, considering factors like whether the input graph is acyclic and has a root node, the types of graphs being processed, and the specific learning goals. (Uwents et al. 2010)

  • Utilise the Kronecker graph model for generating realistic networks due to its ability to create networks that possess all major static network patterns, adhere to temporal evolution patterns, and allow for tractable analysis and rigorous proof. (Leskovec et al. 2008)

  • Use the Gromov-weak topology when studying random metric measure spaces, as it provides a robust and flexible framework for analyzing complex systems involving both geometry and probability. (Greven, Pfaffelhuber, and Winter 2008)

  • Use a hierarchical community detection algorithm that combines modularity optimization with a recursive aggregation process to efficiently analyze large networks, achieving superior performance in terms of computation time and modularity compared to existing methods. (Blondel et al. 2008)

  • Focus on developing a unified technique for deriving Gaussian central limit theorems for linear statistics of eigenvalues in random matrices through the use of second order Poincare inequalities, which provide central limit theorems, rather than traditional approaches that involve hard computations specific to the model in question. (S. Chatterjee 2007)

  • Use dynamic frequent subgraph mining to identify patterns and relationships within complex networks over time. (“Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining” 2007)

  • Consider using explainable artificial intelligence (AI) algorithms, specifically those that generate a distribution of explanations rather than just a single maximum reward sample, to better understand the underlying relationships and potential errors in deep learning models. (Tiejun Cheng et al. 2007)

  • Use a systematic framework for learning from a finite set represented as a graph, developing discrete analogues of various differential operators and constructing a discrete analogue of classical regularization theory based on them, allowing for a wider range of regularization options beyond just the commonly used graph Laplacian-based ones. (“Semi-Supervised Learning” 2006)

  • Consider using the Decision Tree Chunkingless Graph-Based Induction (DT-ClGBI) algorithm for constructing decision trees for graph-structured data, as it enables simultaneous construction of attributes useful for classification through the internal use of the Chunkingless Graph-Based Induction (Cl-GBI) algorithm. (“Advances in Knowledge Discovery and Data Mining” 2006)

  • Carefully consider the appropriate choice of graph neural network architecture when working with unstructured, non-Euclidean data, taking into account factors such as the nature of the data, the specific task at hand, and the available computational resources. (G. E. Hinton and Salakhutdinov 2006)

  • Carefully consider the impact of adding small rank deterministic matrices to large Hermitian random matrices, as this can significantly alter the limiting properties of the spectrum as the size of the matrix increases. (Péché 2005)

  • Utilise a combination of graph spectral measures and Wasserstein distances to identify super-classes within graph neural networks, allowing for improved classification performance in few-shot learning situations. (Borgwardt et al. 2005)

  • Use lifted multicuts to optimize graph decompositions, enabling them to effectively compare clusters and identify optimal decompositions based on minimum cost lifted multicuts. (N. Bansal, Blum, and Chawla 2004)

  • Carefully consider the relationship between random quadrangulations and Aldous Integrated SuperBrownian Excursion (ISE) in order to better understand global properties of distances within random maps.’ (Chassaing and Schaeffer 2003)

  • Focus on identifying and analyzing the sharp metastability threshold for two-dimensional bootstrap percolation processes, which occurs at a specific critical value of π²/18, to accurately predict the behavior of these complex systems. (Holroyd 2003)

  • Utilise Markov Chain Monte Carlo (MCMC) techniques to generate uniformly distributed random Bayesian networks, which can help ensure unbiased testing of inference and learning algorithms, and facilitate empirical investigation of network properties. (“Advances in Knowledge Discovery and Data Mining” 2002)

  • Utilize non-intersecting paths, random tilings, and random matrices to derive measures and establish asymptotic results for various models, including the Aztec diamond, rhombus tilings of an abc-hexagon, a dimer model on a cylindrical brick lattice, and a growth model. (K. Johansson 2002)

  • Carefully consider the impact of boundary effects on the long-range structure of your data when analyzing complex systems, as these effects can significantly influence the overall behavior and outcomes of the system. (Kenyon 2000)

  • Utilize Hammersleys process, a continuous-space interacting particle process, to understand the hydrodynamical limit theorem for a specific type of random Young tableau, thereby proving ELn ~ 2n^(1/2). (Aldous and Diaconis 1995)

  • Ensure that your statistical models accurately capture the underlying dynamics of the system being studied, particularly when dealing with complex phenomena like random dynamical systems. (Crauel and Flandoli 1994)

  • Use a probabilistic generative model-level explanation method like GNNInterpreter to understand the high-level decision-making processes of Graph Neural Networks (GNNs) and identify potential model pitfalls, particularly in critical fields where incorrect predictions could lead to significant negative outcomes. (Debnath et al. 1991)

  • Utilize a combination of Voronoi vertices and Delaunay triangulations to accurately reconstruct surfaces from scattered sample points, providing a provably correct algorithm for this process. (NA?)

  • Utilise a two-phase greedy algorithm for learning graphical models from data, which involves first adding dependencies until reaching a local maximum, followed by deleting dependencies until another local maximum is achieved. (NA?)

  • Consider using graph-based hashing methods, specifically Anchor Graph Hashing (AGH), to efficiently discover the neighborhood structure inherent in your data and learn appropriate compact codes for accurate nearest neighbor searches. (NA?)

  • Consider using a kernel function that follows the syntactic structure of the data, as defined by its type signature in a higher-order logic, when working with structured data. (NA?)

  • Consider incorporating randomization into your graph structures when working with semi-supervised learning techniques, as it can lead to improved accuracy and coverage tradeoffs, particularly when the graph supports small cuts. (NA?)

  • Consider the temporal and sequential aspects of event-based network data when developing prediction and ranking algorithms, rather than focusing solely on static network structures. (NA?)

  • Utilise graph kernels for analysing chemical compound data, specifically focusing on the Tanimoto, MinMax, and Hybrid kernels introduced here, as they offer superior performance compared to previous methods in predicting mutagenicity, toxicity, and anti-cancer activity. (NA?)

  • Consider using the Maximum Relevance Minimum Redundancy (MRMR) principle for gene selection in supervised gene selection procedures, as it provides an optimal pairwise approximation of the conditional mutual information between any two genes given the set of selected variables. (NA?)

  • Utilise the Kronecker graphs’ model for generating synthetic graphs that closely mirror the properties of real graphs, and employ the ‘KronFit’ algorithm for efficiently fitting the model to real networks. (NA?)

  • Consider employing either iterative classification or Gibbs sampling for your collective classification tasks, depending on the specific requirements of your project. (NA?)

  • Consider using the R/Bioconductor package minet (version 1.1.6) for inferring mutual information networks from microarray datasets, offering multiple entropy estimators and inference methods such as relevance networks, ARACNE, CLR, and MRNET, along with integrated accuracy assessment tools like F-scores, PR-curves, and ROC-curves for comparison against reference networks. (NA?)

  • Utilize a standardized graph data set repository for benchmarking purposes in graph-based machine learning, allowing for improved comparison and evaluation of different approaches. (NA?)

  • Utilise a kernel approach to unify vector-based and graph-based semi-supervised clustering methods, enabling them to handle both types of data effectively while improving clustering results through the incorporation of prior information about clusters. (NA?)

  • Focus on developing and testing novel graph kernels based on subtree features, rather than relying solely on traditional walk-based graph kernels, as these new kernels can lead to significant improvements in performance across various application areas. (NA?)

  • Utilize a hierarchical random graph model to identify and understand the hierarchical structure inherent in complex networks, enabling accurate prediction of missing connections and providing valuable insights into various network phenomena. (NA?)

  • Use a novel “co-regularization RKHS” to simplify the analysis of complex multi-view learning problems, leading to significant improvements in both theoretical understanding and practical applications. (NA?)

  • Use the TESLA method for estimating time-varying networks, as it provides a practical and scalable way to recover large-scale networks underlying various sociocultural and biological processes at any desired temporal resolution. (NA?)

  • Utilise a graph-based regularisation framework, specifically GNetMine, to accurately classify heterogeneous information networks. This approach allows for explicit consideration of type differences in links and objects, thereby enabling better organisation of typed information and improved overall classification accuracy. (NA?)

  • Utilize a novel learnable proximity measure for relational retrieval tasks, which employs a weighted combination of simple “path experts,” each corresponding to a particular sequence of labeled edges, rather than the conventional Random Walk with Restart (RWR) model. (NA?)

  • Use a dynamic stochastic block model (DSBM) for analyzing dynamic communities in social networks, which enables simultaneous modeling of communities and your evolution, explicit modeling of community member transitions, and Bayesian inference for estimating uncertain parameters. (NA?)

  • Differentiate various types of connections in building a discriminative classifier to improve classification performance in heterogeneous social media networks. (NA?)

  • Consider using the Gromov-Wasserstein distance when comparing objects, as it offers a more computationally feasible alternative to the Gromov-Hausdorff distance while still maintaining strong theoretical underpinnings. (NA?)

  • Focus on developing scalable and efficient heuristic algorithms for solving the influence maximization problem in large-scale social networks, rather than relying solely on traditional greedy algorithms that struggle with scalability. (NA?)

  • Consider utilising diffusion kernels, a specific class of exponential kernels, for your analysis of graph-based data. These kernels are derived from the heat equation and can be seen as the discretisation of the Gaussian kernel typically applied in Euclidean space. They offer a promising tool for capturing both local and global structure within graph-based datasets, making them particularly useful for tasks such as classification of categorical data. (NA?)

  • Utilize the huge’ R package for high-dimensional undirected graph estimation because it offers several advantages over existing packages such as being written in C, supporting additional model types, offering various functions for data-dependent model selection, data generation, and graph visualization, correcting a minor convergence issue in the graphical lasso algorithm, and allowing for both lossless and lossy screening rules to balance computational and statistical efficiency.’ (NA?)

  • Consider using the GraphLab abstraction for your machine learning and data mining tasks because it enables efficient and scalable asynchronous, dynamic, graph-parallel computation while providing strong data consistency guarantees. (NA?)

  • Explore various combinations of graph construction and label inference algorithms in semi-supervised learning, and consider implementing a novel graph construction method called \(b\)-matching, which outperforms the previously dominant \(k\) nearest neighbor graph method. (NA?)

  • Utilise metadata in conjunction with network structure to enhance the accuracy of community detection in networks, while avoiding the assumption that metadata necessarily correlate with the communities sought. (NA?)

  • Utilize a network integration approach like DTINet to effectively combine heterogeneous data sources for improved drug-target interaction prediction and computational drug repositioning. (NA?)

  • Use persistent homology as a method to analyze topological features of data sets, which involves constructing a filtered simplicial complex based on all possible distances between points in the dataset, and computing the persistence intervals of each feature to identify your significance. (NA?)

  • Focus on understanding and utilizing the concept of invariant and equivariant linear layers when working with graph data, as these layers offer significant advantages in terms of dimensionality reduction and improved performance. (Gori, Monfardini, and Scarselli, n.d.)

  • Utilize activation patterns in the hidden layers of Graph Neural Networks (GNNs) to better understand how GNNs perceive the world, thereby improving the explainability and transparency of these models. (NA?)

  • Use a combination of CiteSpace and Pajek tools to analyze the bibliographic literature from WoS, tracing the temporal evolution and identifying the intellectual structure of your research topic through visual analysis. (NA?)

  • Consider the various ways in which networks evolve over time, including slow evolution, rapid evolution, and streaming networks, and choose appropriate methods for analyzing and maintaining the accuracy of your findings accordingly. (NA?)

  • Adopt the CompGCN framework when working with multi-relational graphs, as it effectively combines various composition operations from knowledge graph embedding techniques to jointly embed both nodes and relations in the graph, thereby improving performance across tasks such as node classification, link prediction, and graph classification. (NA?)

  • Use a graph neural network to analyze the type dependency graph generated through lightweight source code analysis, thereby enabling accurate and efficient type inference for TypeScript programs. (NA?)

  • Use artificial neural networks to study topological phases of matter, as they can accurately and efficiently represent these states in various dimensions. (NA?)

  • Adopt an end-to-end deep learning approach for predicting information cascades, rather than relying solely on hand-crafted features or heuristics. (NA?)

  • Consider using the neural relational inference (NRI) model, which combines a variational auto-encoder with graph neural networks, to effectively learn the dynamics of interacting systems solely from observational data. (NA?)

  • Consider using the SIDE dataset, which uses spatial interpolation techniques to estimate local ethnic diversity, rather than traditional polygon-based datasets that struggle to accurately represent mixed ethnic populations. (NA?)

  • Consider using crystal graph convolutional neural networks (CGCNN) for accurate and interpretable predictions of material properties, as demonstrated by the authors successful application of this technique to predict various properties of crystals with diverse structure types and compositions.’ (NA?)

  • Consider incorporating hierarchical taxonomies into network embeddings to improve performance in tasks such as classification and link prediction, and to address issues of data sparsity. (NA?)

  • Utilise graph convolutional networks when dealing with complex graph structures, as it allows for effective modelling and analysis of the relationships within the graph. (NA?)

  • Prioritize optimizing both memory requirements and computational efficiency when developing graph neural network algorithms, particularly for large-scale graphs. (NA?)

  • Consider implementing hard graph attention operator (hGAO) and channel-wise graph attention operator (cGAO) in your deep learning models for improved performance and reduced computational costs. (NA?)

  • Utilize a combination of relation-aware aggregators, meta-path defined receptive field samplers, and co-attention mechanisms to improve the accuracy and efficiency of recommendation algorithms in social e-commerce settings. (NA?)

  • Employ a semi-supervised learning approach to graph classification problems, using an iterative algorithmic framework that alternately builds or updates classifiers at the instance and hierarchical levels, enforcing consistency between them through a disagreement loss, and carefully selecting high-confidence predicted labels to expand the training set. (NA?)

  • Consider utilizing descriptor-based models for molecular property prediction, as they often provide better prediction accuracy and computational efficiency when compared to graph-based models. (NA?)

  • Consider utilizing a Markov-chain model of random walk through a database to calculate similarities between nodes, as this approach effectively captures the increased connection strength between nodes and performs well in comparison to other methods. (NA?)

  • Utilize discourse-based information when attempting to aggregate passage-level clues for solving logical reasoning problems in QA tasks. (NA?)

  • Consider using the PPRGo model when working with large graphs, as it provides an efficient approximation of information diffusion in GNNs, leading to significant speed gains while maintaining state-of-the-art prediction performance. (NA?)

  • Consider using the “Relational Reflection Transformation” technique when working on entity alignment tasks, as it meets the crucial criteria of “relational differentiation” and “dimensional isometry”, leading to improved accuracy over traditional methods. (NA?)

  • Utilise a Graph Neural Network (GNN) approach to generate structure-aware representations for each entity, allowing for inductive predictions on unseen entities, thereby addressing the limitations of traditional entity alignment methods which rely solely on transduction and ignore potentially valuable attribute information. (NA?)

  • Carefully choose among the four proposed categories of graph neural networks - recurrent graph neural networks, convolutional graph neural networks, graph autoencoders, and spatial-temporal graph neural networks - when working with graph data, considering factors such as the nature of the data, the specific task at hand, and the desired outcome. (NA?)

  • Consider utilizing both descriptor-based and graph-based models when developing machine learning models for molecular property prediction, as each approach offers unique advantages and limitations. (NA?)

  • Consider simplifying graph encoders and implementing efficient negative sampling strategies to improve the efficiency and scalability of entity alignment methods. (NA?)

  • Consider using the DivEA framework to efficiently and accurately align entities across large-scale knowledge graphs by improving task division, increasing coverage of potential mappings, and controlling subtask size. (NA?)

  • Adopt a semi-supervised subgraph recovery approach when dealing with anti-money laundering (AML) issues in the banking sector. This involves reconstructing the original subgraph to an anomaly-free subgraph, solving the AML problem on the subgraph level, and leveraging both unsupervised and supervised learning techniques to improve detection rates and overcome challenges associated with sparse and unbalanced data. (NA?)

  • Utilize a combination of heterogeneous region embedding (HRE) and prompt learning techniques to optimize your analysis of urban data. (NA?)

  • Focus on developing a label-aware high-frequency indicator to detect and prune inter-class edges in order to improve the accuracy of graph anomaly detection algorithms. (NA?)

  • Focus on developing a multi-agent reinforcement learning (MARL) framework to simulate the real-world behavior of fraudsters who share different posts, thereby enabling effective targeted attacks on GNN-based fake news detectors. (NA?)

Graph Representation Learning

  • Focus on developing entity-pair embeddings rather than individual entity embeddings when performing knowledge graph alignment tasks. (Fanourakis et al. 2023)

  • Focus on developing a comprehensive understanding of the problem domain, carefully selecting appropriate feature extraction techniques, and evaluating multiple detection methods to ensure accurate and efficient identification of online transaction fraud. (S. Cao et al. 2019)

  • Consider developing efficient and scalable learning algorithms for handling large, attributed multiplex heterogeneous networks, which involve multiple types of nodes connected through multiple types of edges, and each node possessing a set of different attributes. (Cen et al. 2019)

  • Focus on developing inductive network embedding frameworks that jointly preserve both local proximity and structural identities of nodes, allowing for effective handling of unseen nodes or networks. (Junliang Guo, Xu, and Liu 2018)

  • Consider the interdependence of social rank and proximity-based factors in explaining link formation within networks, rather than treating them as separate elements. (Yupeng Gu et al. 2018)

  • Leverage a novel representation learning model to embed each document in a low dimensional vector space, allowing for efficient name disambiguation via hierarchical agglomerative clustering while preserving privacy. (Baichuan Zhang and Hasan 2017)

  • Consider developing a novel framework called MINES (Multi-dImensional Network Embedding with hierarchical Structure) to effectively represent and analyze complex multi-dimensional networks with hierarchical structures. (P. Cui et al. 2017)

  • Consider utilizing the proposed Embedding Propagation (EP) framework for unsupervised learning of graph-structured data, as it demonstrates superior performance compared to various state-of-the-art unsupervised and semi-supervised learning methods while requiring significantly fewer parameters and hyperparameters. (Garcia-Duran and Niepert 2017)

  • Consider using graph-based word representations rather than traditional vector-space embeddings when working with natural language processing tasks, as they offer superior performance on various tasks and capture the inherent hierarchical structure of language more accurately. (Hamilton, Ying, and Leskovec 2017b)

  • Consider utilizing the node2vec algorithm when conducting prediction tasks over nodes and edges in networks, as it offers a scalable and flexible approach to feature learning that can adapt to diverse connectivity patterns and improve overall prediction accuracy. (Grover and Leskovec 2016)

  • Consider utilizing the node2vec algorithm when conducting prediction tasks over nodes and edges in networks, as it offers a scalable and flexible approach to feature learning that can adapt to diverse connectivity patterns and improve overall prediction accuracy. (Grover and Leskovec 2016)

  • Consider using context-aware network embedding (CANE) instead of traditional context-free embedding techniques for improved accuracy in relation modeling, particularly in cases involving complex interactions between vertices. (R. Johnson and Zhang 2014)

Graph Convolutional Networks (Gcn)

  • Consider combining graph-based intention mining with CTR prediction tasks through end-to-end joint training to address issues such as behavior sparsity and weak generalization. (Feng Li et al. 2021)

  • Consider integrating diversification into the matching process using Graph Convolutional Networks (GCN) to optimize both accuracy and diversity in recommendation systems. (Y. Zheng et al. 2021)

  • Consider using a sequential Graph Convolutional Network (GCN) for active learning, as it enables efficient selection of diverse and informative samples by leveraging the relationships between labeled and unlabeled instances in the data. (Caramalau, Bhattarai, and Kim 2020)

  • Focus on developing a novel neural network-based graph technique that jointly considers “Device aggregation” and “Activity aggregation” in heterogeneous graphs to effectively identify malicious accounts in financial systems. (Ziqi Liu et al. 2020)

  • Utilize a combination of graph neural networks and attention mechanisms to accurately differentiate between confusing law articles in order to improve the accuracy of legal judgment predictions. (N. Xu et al. 2020)

  • Consider utilizing a two-stream graph convolutional network (TSGCNet) when working with 3D dental model segmentation tasks. This approach allows for the separate processing of coordinates and normal vectors, leading to the extraction of more discriminative geometric features and ultimately improving the segmentation performance. (Lingming Zhang et al. 2020)

  • Incorporate relation-awareness into your neighborhood matching models for entity alignment, allowing for enhanced accuracy through the simultaneous consideration of both entities and your connecting relations. (Yao Zhu et al. 2020)

  • Employ a combination of graph convolutional networks (GCN) and neural random forests (NRF) to effectively address the challenges of robust rating prediction and fraudster detection in online review platforms. (Shijie Zhang et al. 2020)

  • Incorporate graph convolutional networks (GCNs) into your analyses to improve the accuracy of detecting illicit transactions in cryptocurrency systems, while also considering the temporal dynamics of the system through the use of EvolveGCN. (Weber et al. 2019)

  • Consider combining graph-based intention mining with CTR prediction tasks through end-to-end joint training to address issues such as behavior sparsity and weak generalization. (Feng Li et al. 2019)

  • Differentiate between social homophily and social influence in order to better understand the dynamic nature of social effects in recommender systems. (Q. Wu et al. 2019)

  • Use a Graph Inference Learning (GIL) framework to improve the performance of semi-supervised node classification tasks by learning the inference of node labels on graph topology, rather than solely relying on conventional graph convolution techniques. (Abu-El-Haija et al. 2018)

  • Consider utilizing multiple sub-graphs rather than a single graph for relation extraction tasks, as demonstrated through the successful implementation of the C-GCN-MG model. (Devlin et al. 2018)

  • Consider extending graph convolutional networks to handle dependency structures efficiently in parallel, while applying a novel pruning strategy to keep only the most relevant information around the shortest path between two entities, thereby enhancing relation extraction performance. (Yuhao Zhang, Qi, and Manning 2018)

  • Focus on developing efficient, localized convolutions for graph convolutional networks (GCNs) by sampling the neighborhood around a node and dynamically constructing a computation graph from this sampled neighborhood, rather than operating on the entire graph during training. (Ying et al. 2018)

  • Explore the use of dependency-based convolutional neural networks and entity mention-based pooling methods for improving event detection accuracy in natural language processing tasks. (Eriguchi, Tsuruoka, and Cho 2017)

  • Consider integrating a Symbolic Graph Reasoning (SGR) layer into your convolutional neural networks (ConvNets) to improve overall performance, particularly in tasks involving large-scale category segmentation and image classification. (Caesar, Uijlings, and Ferrari 2016)

  • Use Graph Convolutional Neural Networks (GCNNs) to effectively model complex correlations among edge weights in a road network, enabling accurate estimation of stochastic weights for edges without data. (Defferrard, Bresson, and Vandergheynst 2016)

  • Utilize a scalable approach for semi-supervised learning on graph-structured data based on an efficient variant of convolutional neural networks operating directly on graphs, which enables better encoding of local graph structure and node features while also improving classification accuracy and efficiency. (Kipf and Welling 2016a)

  • Consider extending the ideas from [2 (Henaff, Bruna, and LeCun 2015)

  • Consider utilizing graph convolutional networks (GCNs) to integrate both semantic and structural information for improved controversy detection in social media posts. (Diederik P. Kingma and Ba 2014)

  • Utilise graph convolutional networks (GCNs) to encode constituent structures and inform a semantic role labelling (SRL) system, as opposed to relying solely on dependency representations of syntax. (Graves 2013)

  • Consider employing a combination of temporal edges representation, edge2node, and structural enhancement techniques when developing a network representation framework for detecting phishing addresses in Ethereum transaction networks. (NA?)

Graph Attention Networks (Gat)

  • Consider implementing a novel labeling strategy when dealing with structured sentiment analysis tasks, specifically by introducing an “essential label set” and a “whole label set”. This approach helps to address the issue of label imbalance and improves overall model performance. (Wenxuan Shi et al. 2022)

  • Consider using graph attention networks (GATs) for handling graph-structured data, as they effectively address the limitations of traditional graph convolutions and provide significant improvements in accuracy and efficiency. (Veličković et al. 2017)

  • Consider using graph attention networks (GATs) for handling graph-structured data, as they effectively address the limitations of traditional graph convolutions and provide significant improvements in accuracy and efficiency. (Sperduti and Starita 1997)

Graph Embedding Techniques

  • Combine both group-level and individual-level anomaly indicators to improve the accuracy of detecting fake reviewer groups in online review systems. (C. Cao et al. 2021)

  • Utilize a supervised borrowing method called SuperBorrow’ to enhance the link prediction performance of multiple widely-used prior KGE methods like TransE, DistMult, ComplEx, and RotatE. This method involves learning to score the suitability of an LDP to represent a without-mention entity pair using pre-trained entity embeddings and contextualized LDP representations.’ (Rezayi et al. 2021)

  • Use a multi-scale contrastive learning approach to improve graph anomaly detection by capturing anomalous patterns across different scales. (M. Jin et al. 2021)

  • Utilize a Heterogeneous Information Network (HIN) model for social review platforms, allowing for a unique representation for each component and performing graph inductive learning on the review data through aggregating features of nearby nodes. This addresses the camouflage issue (fraudsters with genuine reviews) which is often coupled with cold-start, i.e., new fraudsters with genuine first reviews. Additionally, the authors suggest a multi-component classification approach, enabling the (Shehnepoor et al. 2020)

  • Carefully select appropriate sub-sampling strategies when constructing knowledge graph datasets for link prediction tasks, taking into consideration factors like minimum relation frequencies, node degrees, and removing inverse relations to ensure high-quality and interpretable evaluations. (Siddhant Arora 2020)

  • Utilize the insights from word embeddings to better understand and improve the representation of knowledge graph relations. (C. Allen, Balažević, and Hospedales 2019)

  • Consider using the RotatE model for knowledge graph embedding tasks because it is scalable, able to model and infer various relation patterns, and significantly outperforms existing state-of-the-art models for link prediction. (Zhiqing Sun et al. 2019)

  • Carefully consider the potential impact of node polysemy on network embedding models, and explore methods like polysemous deepwalk to improve the accuracy and interpretability of these models. (N. Liu et al. 2019)

  • Incorporate crossover interactions in Knowledge Graph Embedding (KGE) models by learning an interaction matrix to generate multiple specific interaction embeddings, allowing for more accurate and reliable predictions and explanations. (Wen Zhang et al. 2019)

  • Consider utilizing network embedding methods to efficiently analyze large-scale networks, particularly those with evolving structures, limited labeled data, and complex community structures. (Weijie Chen et al. 2018)

  • Utilise hyperbolic spaces rather than Euclidean ones for modelling hierarchical relationships, as hyperbolic spaces offer superior capacity and avoid the limitations associated with Euclidean spaces, such as limited capacity and poor ability to reflect the original tree metric. (Ganea, Bécigneul, and Hofmann 2018)

  • Carefully consider the choice of property to be preserved, scalability, and dimensionality of the embedding when selecting a graph embedding technique for your specific application. (Palash Goyal and Ferrara 2018)

  • Integrate the Hawkes process into network embedding to effectively capture the influence of historical neighbors on current neighbors, thereby improving the accuracy of predictions in various tasks like node classification, link prediction, and embedding visualization. (Zuo et al. 2018)

  • Consider leveraging multiple sources of information when making predictions about sentiment links in online social networks, such as social relationships, profile knowledge, and sentiment extracted through entity-level sentiment extraction techniques. (Hongwei Wang et al. 2018)

  • Consider using feature hashing techniques when dealing with large-scale graph data, as it allows for efficient dimensionality reduction while preserving the inner product between vectors, ultimately improving the overall performance of network representation learning algorithms. (Qixiang Wang et al. 2018)

  • Incorporate adversarial learning principles into your network embedding frameworks to ensure robust and accurate representations. (Q. Dai et al. 2017)

  • Consider using the GeomE framework for knowledge graph embedding tasks, as it offers improved performance over existing methods such as ComplEx, pRotatE, and QuatE, through its ability to capture various relation patterns, provide rich expressiveness, and offer better generalization capacity. (Kadlec, Bajgar, and Kleindienst 2017)

  • Utilize the SimplE algorithm for link prediction tasks in knowledge graphs due to its ability to encode background knowledge, interpretability, and superior performance compared to other tensor factorization methods. (Abadi et al. 2016)

  • Carefully consider the choice of context when performing unsupervised representation learning on graph substructures, as incorrect assumptions about context can lead to lower quality embeddings and reduced classification/clustering accuracy. (Narayanan et al. 2016)

  • Consider using a manifold-based embedding principle (ManifoldE) instead of traditional translation-based principles for knowledge graph embedding, as it addresses issues of ill-posed algebraic systems and overly strict geometric forms, leading to improved accuracy and efficiency in precise link prediction tasks. (Han Xiao, Huang, and Zhu 2015)

  • Develop a novel metric on connected bipartite graphs to effectively distinguish between graphs with the same structure, thereby improving the accuracy of detecting suspicious subgraphs in various domains. (Akoglu, Tong, and Koutra 2014)

  • Utilise a novel node embedding of directed graphs to statistical manifolds, which is based on a global minimisation of pairwise relative entropy and graph geodesics in a non-linear manner. (Diederik P. Kingma and Ba 2014)

  • Consider utilizing both labeled and unlabeled data in feature selection processes, particularly when dealing with limited labeled data, by implementing a semi-supervised feature selection algorithm based on spectral analysis. (Z. Zhao and Liu 2007)

  • Optimize a carefully designed objective function that preserves both the local and global network structures, while utilizing an edge-sampling algorithm to improve the effectiveness and efficiency of the inference. (NA?)

  • Utilise the HIN2Vec framework when exploring meta-paths in heterogeneous information networks for representation learning. (NA?)

  • Carefully consider the heterogeneity of your networks when applying network embedding techniques, and they may need to modify or adapt standard approaches like skip-gram to better account for the complex interactions between different types of nodes. (NA?)

  • Consider utilizing the DeepGL framework when working with graph representation learning tasks, as it offers several benefits such as being inductive, space-efficient, fast, scalable, hierarchical, and interpretable. (NA?)

  • Utilize the REGAL algorithm for efficient and accurate graph alignment, which involves learning node embeddings via xNetMF, a novel matrix factorization technique that preserves structural similarities across multiple networks without requiring explicit construction of a full similarity matrix. (NA?)

  • Develop a graph-level embedding method capable of preserving both local and global network structures to effectively differentiate normal and anomalous subgraphs. (NA?)

  • Consider combining attribute-based information and graph-based information to improve the accuracy and robustness of fraud detection systems, while addressing challenges such as constructing appropriate graphs, developing efficient graph learning methods, and effectively fusing different types of information. (NA?)

  • Utilize a combination of parameterized random walk strategy and hyperbolic Skip-gram model to effectively capture the semantic information and hierarchical structure of tags within a node-tag hybrid network. (NA?)

Graph Neural Network Architectures

  • Use a combination of group aggregation and learnable encodings to improve the accuracy of your fraud detection systems, particularly in cases of low homophily. (“Orion-Wyc/GAGA: Release for the WWW23 Artifacts Available.” 2023)

  • Focus on addressing the challenges of fraudster camouflage and class-imbalance problems in fraud detection using techniques such as reinforcement learning for selecting the most informative neighbors for GNNs, while also considering the use of public datasets for empirical validation. (S. Ghosh et al. 2023)

  • Utilize a combination of reinforcement learning and self-supervision to dynamically adjust the depth and breadth of your graph neural network architecture during the course of training, allowing them to balance computational efficiency against the need to capture complex, multi-layered relationships in your data. (Yingguang Yang et al. 2023)

  • Incorporate cost-sensitive learning into graph neural networks to efficiently address the graph imbalance problem in telecom fraud detection. (X. Hu et al. 2023)

  • Consider using a confidence graph constraint within a consensus graph learning model to improve the quality of the learned consensus graph and enhance the overall performance of incomplete multi-view clustering tasks. (C. Liu et al. 2023)

  • Focus on developing methods that can effectively integrate both intra- and inter-graph information when working on graph-level anomaly detection problems. (Xiaoxiao Ma et al. 2023)

  • Consider incorporating higher-order graph structures through network motifs in order to improve the accuracy of financial default predictions. (Daixin Wang et al. 2023)

  • Consider using a prompt-based fine-tuning framework on graph neural networks (GNNs) for voucher abuse detection, as it effectively bridges the gap between GNN pre-training and downstream tasks by reformulating the downstream node classification into a similar form as the pretext task in pre-training. (Z. Wen et al. 2023)

  • Utilize path-based explanations for heterogeneous link prediction tasks due to your superior interpretability, scalability, and ability to handle graph heterogeneity. (Shichang Zhang et al. 2023)

  • Carefully consider the challenges posed by big data, high dimensionality, heterogeneity, lack of labeled data, unbalanced data, unclean data, absence of standardized metrics and benchmarks, changing nature of anomalies, dispersion of sources, obscurity of some anomalies, and the need for robust and scalable algorithms when developing graph-based deep learning methods for anomaly detection in distributed systems. (Pazho et al. 2022)

  • Consider utilizing a user-entity graph that employs edge types and attributes alongside node attributes to capture more behavior than previously considered, and subsequently develop a graph neural network model that takes into account neighborhood, edge types, and attributes for distinguishing spammers from non-spammers. (P. Agarwal et al. 2022)

  • Simultaneously model the homophilic and heterophilic connections in a fraud graph, allowing them to differentiate between nodes with different labels and improve the accuracy of fraud detection. (F. Shi et al. 2022)

  • Consider utilizing an adaptive multi-frequency graph neural network (AMNet) for anomaly detection tasks, as it allows for the simultaneous capture of both low-frequency and high-frequency signals, enabling better distinction between normal and anomalous nodes. (Z. Chai et al. 2022)

  • Consider incorporating user preference analysis alongside traditional content and context analysis in order to improve the accuracy of fake news detection systems. (Y. Dou et al. 2021)

  • Consider utilizing semi-supervised learning techniques, specifically those based on graph representations, to improve classification performance by leveraging both labeled and unlabeled data. (Z. Song et al. 2021)

  • Utilize Graph Neural Networks (GNN) for inductive text classification, building individual graphs for each document to capture contextual word relationships and generate effective embeddings for unseen words. This approach was found to outperform state-of-the-art text classification methods across multiple benchmark datasets. (F. Hu et al. 2021)

  • Integrate the anomaly detection process and GNN learning process into a unified framework called RARE-GNN, allowing them to mitigate the adverse effects of anomalies and learn powerful GNNs. (K. Ding, Shan, and Liu 2021)

  • Consider using the Eland framework, which combines action sequence augmentation and graph anomaly detection, to improve the accuracy of early-stage graph anomaly detection when data is limited. (Tong Zhao et al. 2021)

  • Consider using a dual-tier heterogeneous graph (DHG) model for document-level relation extraction, which effectively separates document modeling from multi-hop reasoning, leading to improved accuracy and better handling of complex structures within texts. (Zhenyu Zhang et al. 2020)

  • Consider utilizing graph representation and learning frameworks like FANG to improve the accuracy and efficiency of detecting fake news by incorporating social context and engagement patterns. (V.-H. Nguyen et al. 2020)

  • Develop a novel framework called “Adaptive Target-Behavior Relational Graph” (ATBRG) to effectively capture structural relations of target user-item pairs over a knowledge graph (KG) using graph connect and graph prune techniques, and to fully distill structural information from the sub-graph connected by rich relations in an end-to-end fashion using a relation-aware extractor layer and representation activation layer. (Yufei Feng et al. 2020)

  • Consider and address the potential issues arising from context, feature, and relation inconsistencies when applying graph neural networks (GNNs) to fraud detection tasks. (Zhiwei Liu et al. 2020)

  • Consider integrating graph neural networks (GNNs) into your encoder-decoder parsers to effectively encode the structure of database (DB) schemas, thereby significantly improving the accuracy of text-to-SQL parsing tasks. (Bogin, Gardner, and Berant 2019)

  • Adopt an edge-oriented graph neural model for document-level relation extraction, which uses different types of nodes and edges to create a document-level graph and employs an inference mechanism on the graph edges to enable learning of both intra- and inter-sentence relations using multi-instance learning. (Christopoulou, Miwa, and Ananiadou 2019)

  • Consider implementing a Kernel Graph Attention Network (KGAT) for more fine-grained fact verification, as it effectively measures the importance of evidence nodes and conducts precise evidence propagation within the graph, leading to improved accuracy in identifying true versus false claims. (Zhenghao Liu et al. 2019)

  • Utilize regularized supervised machine learning algorithms for estimating node importance in Knowledge Graphs (KGs), as opposed to non-trainable solutions, and specifically, they should consider using Graph Neural Networks (GNNs) like GENI, which addresses the unique challenges associated with supervised estimation of node importance in KGs. (N. Park et al. 2019)

  • Consider using the SEED (Sampling, Encoding, and Embedding Distributions) framework for inductive and unsupervised representation learning on graph structured objects, as it enables efficient encoding of subgraphs and measures graph similarity through distribution distance between collections of subgraph vectors. (Arjovsky, Chintala, and Bottou 2017)

  • Utilise an Entity-based Narrative Graph (ENG) to effectively model the internal-states of characters in a story, allowing for accurate predictions of character mental states and desire fulfillment. (Yoon Kim 2014)

  • Consider using the Capsule Graph Neural Network (CapsGNN) approach for generating high-quality graph embeddings, as it effectively addresses limitations in current GNN-based graph embeddings algorithms by leveraging the concept of capsules and dynamic routing mechanisms to better capture and represent graph properties. (Bruna et al. 2013)

  • Pre-train Graph Neural Networks (GNNs) at both the node and graph levels to improve generalization and avoid negative transfer across downstream tasks. (Bemis and Murcko 1996)

Computational Biology And Bioinformatics

  • Consider combining neural networks with sequence representations obtained from protein language models to enhance the ability to detect remote homologues within protein domain classification systems. (Nallapareddy et al. 2023)

  • Utilise a combination of phylogenomic principles and automated procedures like RIO (Resampled Inference of Orthologs) to accurately determine orthologs and paralogs in gene trees, thereby improving the accuracy of functional annotation transfer. (Faltejsková and Vondrášek 2023)

  • Utilize MSA Transformer, a protein language model, for generating novel protein sequences through an iterative masking process, as it demonstrates superior performance in terms of homology, coevolution, and structural scores compared to traditional Potts models like bmDCA. (Sgarbossa, Lupo, and Bitbol 2023)

  • Use a deep learning-based model called GEARS to predict the gene expression outcome of combinatorially perturbing a set of one or more genes, which can improve the efficiency and effectiveness of future perturbational experiments. (Roohani, Huang, and Leskovec 2023)

  • Consider utilizing a combination of unsupervised protein language modeling, structural context through AlphaFold-derived systems, and fine-tuning on weak labels from population frequency data to achieve state-of-the-art missense pathogenicity predictions without explicit training on such data. (Jun Cheng et al. 2023)

  • Utilise a transformer-based conditional language model trained solely on evolutionary sequence data to create functional artificial proteins across various protein families. (Madani et al. 2023)

  • Consider utilizing large language models (LLMs) like GPT-3 for solving various tasks in chemistry and materials science, as they can provide comparable or superior performance to traditional machine learning models while requiring minimal expertise and data. (A. D. White et al. 2022a)

  • Consider employing a Decomposed Fusion with Soft Prompt (DFSP) framework for Compositional Zero-Shot Learning (CZSL) tasks, which involves integrating vision-language models (VLMs) to enhance the understanding of unseen compositions and bridge the domain gap between seen and unseen sets. (S. H. Bach et al. 2022)

  • Focus on developing large-scale attention-based protein language models trained on extensive protein sequence datasets to achieve state-of-the-art performance in capturing the distribution of observed evolutionary sequences, generating novel viable sequences, and predicting protein fitness without additional finetuning. (Nijkamp et al. 2022)

  • Carefully select appropriate CRISPR-Cas systems and optimize your assays for desired readouts, considering factors like sensitivity, specificity, cost, and ease of implementation, depending on the intended application. (Kaminski et al. 2021)

  • Consider using log transformation to improve the accuracy of dating phylogenies, particularly when dealing with complex models. (Mai and Mirarab 2020)

  • Integrate multiple sources of information, such as RNA-sequencing data and CLIP binding data, to improve the accuracy of microRNA target prediction models. (Weijun Liu and Wang 2019)

  • Utilise a novel reinforcement learning (RL) formulation for molecular design in Cartesian coordinates, enabling the extension of molecular design to a wider range of molecules and the application of reward functions based on fundamental physical properties such as energy. (Nathan Brown et al. 2019)

  • Utilise a combination of graph neural networks and reinforcement learning to develop a Graph Transformation Policy Network (GTPN) for accurately predicting chemical reactions without reliance on handcrafted or heuristically extracted reaction rules/templates. (K. Do, Tran, and Venkatesh 2018)

  • Consider applying deep learning techniques, such as deep neural networks, convolutional neural networks, recurrent neural networks, and emerging architectures, to bioinformatics domains like omics, biomedical imaging, and biomedical signal processing, taking into account theoretical and practical issues like imbalanced data, interpretation, hyperparameter optimization, multimodal deep learning, and training acceleration. (S. Min, Lee, and Yoon 2016)

  • Use techniques such as restriction, compression, and parallelism to manage and analyze large genomic datasets effectively. (Lawrence and Morgan 2014)

  • Critically evaluate convergence in your Markov Chain Monte Carlo (MCMC) analyses through comparison of samples from independent MCMC runs, utilizing tools such as the average standard deviation of split frequencies (ASDSF) to quantify similarity among samples. (Ronquist et al. 2012)

  • Consider incorporating a flexible global trend and a variance model into your Gaussian processes (GP) models to improve accuracy and overcome issues caused by the assumption of second-order stationarity. (S. Ba and Joseph 2012)

  • Carefully choose appropriate semantic similarity measures and mixing strategies when analyzing protein-protein interactions, taking into consideration factors like ontology type, reliability of annotations, and potential biases in the calculation of semantic similarity. (Guzzi et al. 2011)

  • Adopt a piecewise stationary model for genomic features, allowing for more accurate and robust statistical inference in genomics. (Bickel et al. 2010)

  • Use Lasso, a linear regression algorithm, to simplify complex models like the Kaplan model for predicting nucleosome occupancy, as it allows for faster model generation, subset selection, and easier interpretation of results. (Tillo and Hughes 2009)

  • Carefully evaluate and choose among the various semantic similarity measures available for biomedical ontologies, taking into consideration factors such as scope, data source, metric, and applicability to your specific research question. (Pesquita et al. 2009)

  • Carefully consider the choice of corpus when comparing the performance of PPI extraction methods, as the choice of corpus can have a larger impact on the result than the choice between a naive and an advanced PPI extraction method. (Pyysalo et al. 2008)

  • Consider integrating various types of data, such as text and protein sequence data, to achieve higher accuracy in predicting protein subcellular localization. (Shatkay et al. 2007)

  • Utilise a model-based background adjustment for oligonucleotide expression arrays to improve the accuracy of your results. (Zhijin Wu et al. 2004)

  • Utilize multiple search algorithms to analyze tandem mass spectra, especially in situations where the spectra are of poorer quality or exhibit unusual fragmentation patterns, as different algorithms demonstrate varying degrees of selectivity and sensitivity. (Sadygov, Cociorva, and Yates 2004)

  • Carefully consider the order of limits taken in your analyses, particularly when studying phenomena like aging, as this can impact the interpretation of results and conclusions drawn. (Arous, Dembo, and Guionnet 2001)

  • Consider using implicit fitness sharing in your genetic algorithms to promote diversity and cooperation within populations, allowing for better adaptation to changing environments and improved performance in complex tasks. (NA?)

  • Use a support vector machine (SVM) learning system to predict protein-protein interactions based solely on primary structure and associated physicochemical properties, achieving an impressive 80% inductive accuracy rate. (NA?)

  • Carefully integrate diverse data sources, such as gene expression and phylogenetic profiles, through methods like early, intermediate, and late integration, to improve the predictive power of support vector machines in identifying gene functions. (NA?)

  • Use a leave-one-out cross-validation testing method to prevent over-estimation of prediction accuracy when training and evaluating gene-expression-based models for cancer classification. (NA?)

  • Utilize machine learning algorithms to enhance the accuracy and coverage of subcellular localization predictions, thereby improving the understanding of protein functions and facilitating your purification. (NA?)

  • Use the Phospho.ELM database to improve your understanding of protein kinase substrate specificity and enhance the accuracy of phosphorylation site predictions through the application of machine learning methods trained on high-quality, experimentally validated data. (NA?)

  • Consider utilizing data mining and machine learning techniques, particularly the MOLFEA algorithm for generating descriptors and the rule learner PART or support vector machines for inducing structure-activity relationships (SARs) from these descriptors, to achieve improved predictive accuracies in identifying mutagenicity-inducing substructures and SARs of noncongeneric compounds. (NA?)

  • Consider utilizing intensity-based protein identification techniques when working with tandem mass spectrometry (MS/MS) data, as it significantly improves peptide and protein identification accuracy while maintaining sensitivity. (NA?)

  • Utilize maximum-entropy techniques, particularly sequential-update algorithms, for modeling species geographic distributions due to your ability to handle a vast number of features effectively. (NA?)

  • Consider utilizing a diverse range of machine learning methods, such as Raplex, boosted wrapper induction, memory-based learning, transformation-based learning, support vector machines, and maximum entropy, to improve the performance of information extraction systems for detecting human protein names and your interactions within Medline abstracts. (NA?)

  • Employ a combination of local contiguous structural and sequence information to accurately differentiate real pre-miRNAs from pseudo pre-miRNAs, thereby improving the efficiency of microRNA detection and prediction. (NA?)

  • Consider utilizing an artificial immune system based on the clonal selection principle to effectively solve multiobjective optimization problems, as demonstrated through successful comparisons with existing state-of-the-art algorithms. (NA?)

  • Consider using machine learning algorithms like Support Vector Machines (SVM) to distinguish enzyme structures from non-enzymes without relying on alignments, achieving up to 80% accuracy through a combination of 36 carefully selected features. (NA?)

  • Avoid using non-co-localized negative examples when evaluating the accuracy of a protein-protein interaction classifier, as doing so introduces bias and leads to overly optimistic estimates of classifier performance. (NA?)

  • Utilize sophisticated bioinformatics tools like VANTED to effectively analyze and visualize large-scale biochemical datasets in the context of biological networks, thereby helping them deduce biologically meaningful interpretations and gain deeper understanding of biological processes. (NA?)

  • Prioritize collecting biologically relevant data sets for building accurate machine learning models, as opposed to relying solely on computationally derived negative examples. (NA?)

  • Focus on developing robust statistical methods capable of handling high-dimensional data with limited sample sizes, while carefully considering potential sources of bias and noise during validation processes. (NA?)

  • Consider using Maxent, a general-purpose method for making predictions or inferences from incomplete information, for presence-only modeling of species distributions, as it offers numerous advantages including requiring only presence data, handling both continuous and categorical data, and providing a mathematically concise definition. (NA?)

  • Utilise a combination of machine learning strategies, particularly Relevance Vector Machines (RVM) and Bayesian statistical methods, to effectively analyse gene expression data and promoter sequences in order to establish accurate transcriptional regulatory networks. (NA?)

  • Utilise the concept of conservation of momentum’ in order to effectively analyse and model the evolution of shapes through the use of geodesic shooting techniques. (NA?)

  • Consider employing multiple complementary methods, such as the profile and stability methods, to improve the accuracy and robustness of your analyses in identifying deleterious single nucleotide polymorphisms (SNPs) in the human population. (NA?)

  • Utilise a combination of high-throughput and low-throughput experimental data to create a robust, high-quality mammalian protein-protein interaction network, which can subsequently be used to generate subnetworks from lists of mammalian genes or proteins, thereby providing insight into the underlying mechanisms of cellular processes. (NA?)

  • Consider integrating multiple types of data, such as protein-protein interactions and Gene Ontology (GO) annotations, to increase the reliability of protein-protein interaction networks and improve the identification of functional modules within them. (NA?)

  • Consider developing SVM models using PSSM profiles for predicting DNA-binding proteins, as it improves the accuracy by 6-7%, while taking care to ensure the quality of PSSM profiles by generating them from similar sequences. (NA?)

  • Consider utilizing multiple classification techniques, such as Quantitative Matrix (QM), Artificial Neural Network (ANN), and Support Vector Machine (SVM), to accurately predict antibacterial peptides based on your amino acid composition and residue preferences at different positions. (NA?)

  • Consider combining both direct and indirect biological data sources in a supervised learning framework when predicting protein-protein interactions, as this approach may reduce false positive or negative results. (NA?)

  • Utilize the PHOSIDA tool to analyze the structure and evolution of the phosphoprotome, taking advantage of its ability to provide secondary structure and accessibility information for each phosphosite, evaluate evolutionary constraints on the phosphoproteome, and generate predictions for phosphorylation sites using a support vector machine. (NA?)

  • Consider incorporating time-series expression data along with ChIP-chip or motif data to create a comprehensive understanding of the dynamic regulatory networks within cells. (NA?)

  • Consider integrating formal and algorithmic approaches into biological research to transform biology into a more precise engineering discipline, thereby unlocking the full potential of executable biology as a mainstream biological technique. (NA?)

  • Utilize a combination of experimental methods and computational tools to accurately predict the subcellular localization of proteins, improving the quality of high-throughput data and facilitating the identification of compartment-specific protein complexes and networks. (NA?)

  • Consider combining multiple relevant features and utilizing advanced machine learning algorithms like random forest to enhance the accuracy and robustness of your models in tasks such as distinguishing real pre-miRNAs from pseudo ones. (NA?)

  • Use a diverse range of expression profiles to maximize the recall and precision of your network inference algorithms, while also considering the limitations of motif analysis in cases of combinatorial or conditional regulation. (NA?)

  • Utilise a machine learning method based on a support vector machine (SVM) combined with a kernel function and a conjoint triad feature abstract for the prediction of Protein-Protein Interactions (PPIs) based solely on the primary sequences of proteins. This approach reduces the risk of overfitting and allows for the effective reproduction of various levels of networks of PPIs, making it a valuable tool for exploring networks for newly discovered proteins with unknown bi (NA?)

  • Carefully consider the choice of evaluation strategy, corpus, and applied metrics when comparing the performance of different machine learning approaches for protein-protein interaction extraction, as these factors significantly affect the interpretation of results. (NA?)

  • Consider utilizing Random Forest Algorithm for predicting glycosylation sites in protein sequences, as it offers numerous advantages such as handling mixed data types, preventing overfitting, and maintaining performance despite noise in the data. (NA?)

  • Focus on improving the accuracy of HLA-II binding predictions for large-scale studies, such as proteome-wide epitope mapping, as currently available servers offer only limited prediction accuracy. (NA?)

  • Consider using positional scanning combinatorial peptide libraries for characterizing the binding specificity of MHC class I molecules due to its cost effectiveness, quantitative nature, and lack of bias. (NA?)

  • Consider utilizing a hierarchical binary decision tree approach for complex multi-tissue classification problems, as it allows for simplified decision making through breaking down the problem into smaller binary choices, while also providing flexibility in structure and choice of classifiers. (NA?)

  • Develop a comprehensive classifier system named “microPred” for distinguishing real pre-miRNA hairpins from both pseudo hairpins and other non-coding RNAs (ncRNAs) by utilizing appropriate machine learning techniques, including the use of a more complete and representative ncRNA and pseudo hairpin dataset, introduction of new biologically relevant features, feature selection, application of class imbalance learning methods, and extensive and systematic training (NA?)

  • Consider developing a database strategy that allows searching based on all potential ionisation products predicted to form during electrospray ionisation (ESI) to improve the efficiency and accuracy of metabolite signal identification in accurate mass metabolomics data. (NA?)

  • Consider utilizing machine learning algorithms like Random Forest for identifying potential epistatic interactions in genome-wide association studies, as they offer a way to effectively reduce the search space for these interactions from an astronomic number of all possible combinations of genetic variants to a manageable set of candidates. (NA?)

  • Develop a generic method for assigning reliability scores to each surface accessibility prediction as an inherent part of the training process, allowing them to identify subsets of highly reliable predictions across all ranges of surface exposure. (NA?)

  • Use a comprehensive dataset of around 6000 non-confidential compounds with known biological activities in the Ames mutagenicity test as a benchmark when comparing various modelling methodologies for predicting Ames mutagenicity. (NA?)

  • Utilise the MetaboAnalyst web server for efficient, accurate, and user-friendly metabolomic data analysis, incorporating a broad array of data types and analytical methods. (NA?)

  • Employ machine learning algorithms to identify and classify effectors in bacterial pathogens, using a diverse set of features covering genomic attributes, evolutionary-based attributes, regulatory network attributes, and attributes specific to the pathogenesis system. (NA?)

  • Utilise the Topological Clustering Semantic Similarity (TCSS) algorithm when computing semantic similarity between Gene Ontology (GO) terms annotated to proteins in interaction datasets. This algorithm takes into account the uneven distribution of biological knowledge representation in various branches of the GO graph, thereby improving the accuracy of protein-protein interaction predictions. (NA?)

  • Avoid using homologous peptides in both training and testing datasets to obtain real-world estimates of prediction performance metrics, while still considering all training data for maximum performance in end-user applications. (NA?)

  • Consider using semi-supervised learning methods, specifically Laplacian regularized least square (LapRLS) and NetLapRLS, to effectively integrate information from chemical, genomic, and drug-protein interaction spaces for accurate drug-protein interaction prediction. (NA?)

  • Use presence-absence data whenever possible, as it provides better information about species prevalence and is less susceptible to sample selection bias compared to presence-only data. (NA?)

  • Consider combining machine learning techniques with hidden Markov modeling to achieve better accuracy when analyzing time-resolved live cell imaging data. (NA?)

  • Utilise a combination of machine learning and decision-theoretic planning to optimise the allocation of human effort in consensus tasks, thereby leveraging the complementary strengths of humans and computer agents to solve these tasks more efficiently. (NA?)

  • Adopt a combinatorial approach to identify subtle signals in DNA sequences, focusing on non-signals (spurious similarities) rather than the signal itself, and utilizing the WINNOWER algorithm to efficiently remove spurious edges and preserve signal edges. (NA?)

  • Focus on identifying the critical points in biological systems, as these points represent a balance between order and disorder, and may hold clues to understanding complex biological processes. (NA?)

  • Consider employing both molecular dynamics simulations and virtual screening methods in tandem to accurately predict and optimize the biological activity of antimicrobial peptides (AMPs) based on your primary amino-acid sequences. (NA?)

  • Consider combining multiple prediction methods through a meta-method approach to improve the accuracy of protein disorder prediction. (NA?)

  • Utilize established benchmark datasets containing cases with known outcomes and appropriate evaluation measures to conduct systematic method performance analysis, allowing for objective and quantitative comparison of prediction tools. (NA?)

  • Utilize machine learning techniques to analyze large amounts of genomic data, specifically focusing on transcription factor binding site patterns, in order to accurately identify and categorize distinct types of genomic regions across multiple cell lines. (NA?)

  • Consider employing a “community approach” to improve the accuracy and robustness of gene regulatory network predictions, especially when dealing with poorly studied organisms or incomplete datasets. (NA?)

  • Carefully choose appropriate simulation tools based on your specific research goals, considering factors like computational efficiency, level of detail needed, and compatibility with existing data sets. (NA?)

  • Utilise machine learning techniques to approximate density functionals, allowing them to achieve greater accuracy in your predictions without requiring extensive insight into the underlying physics. (NA?)

  • Consider applying alignment-independent methods, specifically the auto- and cross-covariance (ACC) transformation, to develop models for allergen recognition based on the main chemical properties of amino acid sequences. (NA?)

  • Consider employing a combination of Principal Component Analysis (PCA) and Ensemble Extreme Learning Machines (ELM) for improved prediction accuracy and reduced computation time in the context of protein-protein interaction (PPI) prediction from amino acid sequences. (NA?)

  • Use a combination of sequence-based and structure-based algorithms to achieve higher overall accuracy in predicting the deleterious effects of human protein variants. (NA?)

  • Consider focusing on mixed selectivity neurons in the prefrontal cortex, as they offer a significant computational advantage over specialized responses in terms of the repertoire of input-output functions implementable by readout neurons, and this advantage originates from the highly diverse nonlinear selectivity to mixtures of task-relevant variables, a signature of high-dimensional neural representations. (NA?)

  • Use multiple machine learning approaches, such as support vector machines, to build accurate models for predicting toxicity of peptides and proteins, taking into account factors such as amino acid composition, dipeptide composition, and motif identification. (NA?)

  • Utilize machine learning models that incorporate both genomic features of cell lines and chemical properties of drugs to effectively predict the response of cancer cell lines to drug treatments, thereby improving the efficiency and accuracy of drug screening processes. (NA?)

  • Consider using the query-centric auto-(k) nearest-neighbor (QCauto) method for DNA barcoding, as it consistently produces low rates of misidentification across various loci and situations, making it a reliable choice for accurate taxonomic identification. (NA?)

  • Focus on the stability of gene selection methods, as opposed to simply evaluating them based on post-selection accuracy, since the former is essential for determining the significance of results. (NA?)

  • Balance model complexity according to study objectives, data attributes, and an understanding of how these interact with underlying biological processes, aiming to avoid both underfitting and overfitting. (NA?)

  • Consider the importance of seed match, conservation, free energy, and site accessibility when selecting and interpreting the output of various miRNA target prediction tools. (NA?)

  • Develop a high-throughput variant prioritization pipeline that integrates various functional annotations, predicts nucleotide-level loss- and gain-of-function events, analyzes evolutionary and human population-level conservation, links variants with target genes using data from the Roadmap Epigenomics Project, incorporates network topology analysis, gene functions, and user annotations to investigate these variant-gene linkages, and identifies recurrent elements from both user-input and publicly (NA?)

  • Utilize ridge regression models for predicting drug responses in patients, as it allows for the inclusion of all genes in the model and performs well across various datasets. (NA?)

  • Utilize a comprehensive approach to modeling RNA-binding protein (RBP) binding preferences, incorporating both sequence and structure information through the use of graph-kernels, rather than relying solely on sequence-based methods. (NA?)

  • Consider incorporating multiple data sources such as drug phenotype, therapeutic, structure, and genome information when developing machine learning models for predicting drug-drug interactions (DDIs), as doing so can potentially enhance the performance of the predictions. (NA?)

  • Consider using the k-nearest neighbor (kNN) algorithm with k = 1 and lag = 5 for accurate allergen prediction, achieving high sensitivity, specificity, positive predictive value, F1 score, and Matthews correlation coefficient. (NA?)

  • Use the open-source tool “Normalyzer” to evaluate the suitability of different normalization methods for your specific dataset, considering both quantitative and qualitative factors, in order to reduce systematic biases and improve the validity of downstream analyses. (NA?)

  • Consider utilizing the Protein Interaction Quantitation (PIQ) method when attempting to accurately predict transcription factor (TF) binding sites from DNase-seq data. (NA?)

  • Utilise multiple sequencing technologies and bioinformatic tools to reduce bias and increase the accuracy of genetic variant identification. (NA?)

  • Utilize an unsupervised spectral approach called Eigen’ for scoring variants, which doesn’t rely on labeled training data, thus avoiding bias or inaccuracies inherent in supervised methods.’ (NA?)

  • Utilize the Similarity Network Fusion (SNF) technique to effectively aggregate diverse types of genome-wide data, thereby creating a comprehensive view of a specific disease or biological process. (NA?)

  • Consider using the CFM-ID web server for automated metabolite identification, as it offers significant improvements in speed and accuracy over manual interpretation of MS/MS spectra and outperforms existing methods like MetFrag and FingerId. (NA?)

  • Combine homology-based inference and machine learning techniques to improve the accuracy of protein sub-cellular localization predictions. (NA?)

  • Carefully select appropriate machine learning algorithms for specific tasks in chemoinformatics, considering factors such as the nature of the data, the desired outcome, and potential limitations of each algorithm. (NA?)

  • Utilize the BAGEL (Bayesian Analysis of Gene Essentiality) computational framework for identifying essential genes from pooled library screens, as it provides increased sensitivity and reduced runtime compared to existing methods. (NA?)

  • Consider implementing a multi-agent machine learning system, providing predicted free energy change values and a corresponding prediction confidence estimation, offering high throughput scanning for multi-point mutations, and having a specific mode for the prediction of stabilizing disulfide bonds when developing a method for predicting changes in stability upon point mutation in proteins. (NA?)

  • Utilize the Ensembl Regulatory Build to create a comprehensive and accurate map of the human genomes regulatory regions, integrating diverse sources of public data on epigenetic marks and transcription factor binding, and verifying it against independent assays for sensitivity.’ (NA?)

  • Explore the use of deep learning networks in conjunction with Sov score evaluation and the inclusion of novel features such as Atchley factors to potentially enhance the accuracy of protein secondary structure prediction. (NA?)

  • Utilise the extended set of functional groups (EFG) when creating quantitative structure-activity relationship (QSAR) models because they offer greater prediction accuracy and easier interpretation compared to the previously used CheckMol-FG set. (NA?)

  • Carefully choose between supervised, unsupervised, and semi-supervised machine learning approaches depending on the availability and similarity of training and test datasets, considering factors such as interpretability, predictive accuracy, and incorporation of prior knowledge. (NA?)

  • Thoroughly evaluate the quality of individual data types, implement appropriate quality control measures, and apply effective data reduction techniques before attempting to integrate diverse datasets for comprehensive biological insights. (NA?)

  • Adopt a continuous distributed representation of biological sequences, named bio-vectors (BioVec), for deep learning applications in proteomics and genomics, as it enables accurate information retrieval about protein structure and can serve as pre-training for various applications. (NA?)

  • Consider integrating 3D DNA shape information into your models of transcription factor binding specificities, as doing so can lead to improved prediction accuracy and reduce the dimensionality of the feature space. (NA?)

  • Use supervised feature selection methods in genomic prediction studies, especially when dealing with traits influenced by loci of moderate effect size, as it allows for more accurate predictions and reduces computational costs. (NA?)

  • Employ a combination of k-gram feature representation calculated as Multivariate Mutual Information (MMI) and normalized Moreau-Broto Autocorrelation (NMBAC) when analyzing protein-protein interaction (PPI) data. (NA?)

  • Utilize advanced computational methods like machine learning to analyze existing IncRNA-disease associations and predict potential ones, thereby contributing to our understanding of complex diseases at the IncRNA level, improving disease biomarker detection, and facilitating better disease diagnosis, treatment, prognosis, and prevention. (NA?)

  • Consider applying deep learning techniques, such as deep neural networks, convolutional neural networks, recurrent neural networks, and emerging architectures, to bioinformatics domains like omics, biomedical imaging, and biomedical signal processing, in order to effectively extract valuable knowledge from big data. (NA?)

  • Utilize deep learning techniques, specifically deep neural networks (DNNs), for toxicity prediction, as they excel in creating task-specific features and outperform previous approaches in various domains. (NA?)

  • Use a combination of experimental annotations and improved methods for function prediction to achieve higher accuracy in protein function prediction. (NA?)

  • Consider adopting a pan-allele, pan-length training pipeline for MHC class I binding prediction, as it combines the benefits of pan-specific training and multi-length binding information, leading to improved predictive performance for alleles covered with limited or no binding data. (NA?)

  • Consider combining multiple techniques and data sources, such as in silico fragmentation, reference and patent information, and suspect screening, to improve the accuracy and reliability of small molecule identification in mass spectrometry. (NA?)

  • Consider using the proposed MLS spike method for extracting neuronal spiking activity from large-scale two-photon recordings, as it outperforms existing state-of-the-art algorithms and enables accurate spike extraction from large-scale recordings. (NA?)

  • Carefully consider and address batch effects in gene expression datasets, as they can account for a substantial proportion of the variance and impact the accuracy of downstream analyses. (NA?)

  • Use a combination of machine learning techniques and a brain-specific functional interaction network to predict autism risk genes, allowing for the identification of novel candidate genes with minimal or no prior genetic evidence. (NA?)

  • Utilize the Regularized Entropy Match (REMatch) approach to effectively combine local descriptors and create a global kernel for describing the similarity of both whole molecular and bulk periodic structures, enabling the navigation of alchemical and structural complexity within a unified framework. (NA?)

  • Utilize a combination of within-score and between-score approaches when analyzing miRNA-disease associations, taking into account both miRNA functional similarity and disease semantic similarity, along with Gaussian interaction profile kernel similarity for diseases and miRNAs. (NA?)

  • Utilize a genetic algorithm approach combined with on-demand property prediction models to efficiently identify and optimize n-block polymers with desired dielectric constants and band gaps, rather than relying solely on enumeration or brute force searches. (NA?)

  • Utilise the SCENIC (Single-Cell rEgulatory Network Inference and Clustering) method to optimally characterise gene regulatory networks using single-cell RNA-seq data, and subsequently improve cellular state identification through the inferred networks. (NA?)

  • Combine rule-based enumeration with machine learning techniques to accurately predict the outcomes of chemical reactions, thereby improving the efficiency and reliability of synthesis planning. (NA?)

  • Use simulated datasets with ground truth information to evaluate the accuracy and reliability of tractography algorithms, while acknowledging the inherent ambiguities in tract reconstruction based on orientation information alone. (NA?)

  • Use the Online Active Set method to Infer Spikes (OASIS) for fast online deconvolution of calcium imaging data, as it provides remarkable increases in processing speed and allows for real-time online estimation of neural activity during the imaging session. (NA?)

  • Utilize Gradient Domain Machine Learning (GDML) to create accurate molecular force fields by learning the functional relationship between atomic coordinates and interatomic forces, rather than computing the gradient of the Potential Energy Surface (PES). (NA?)

  • Consider utilizing the machine learning prediction tool called LOCALIZER, which can effectively predict the subcellular localization of both plant and effector proteins within the plant cell, thereby improving the understanding of pathogen-host interactions. (NA?)

  • Consider employing a deep learning framework, specifically DeepCRISPR, to effectively integrate diverse data sources, optimize sgRNA design, and improve predictions of both on-target and off-target effects in CRISPR applications. (NA?)

  • Utilise the Scanpy toolkit for efficient, modular, and scalable analysis of large-scale single-cell datasets, enabling them to perform various tasks including regression, normalisation, identification of highly variable genes, clustering, and pseudotemporal ordering. (NA?)

  • Utilize the SchNetPack toolbox for developing and implementing deep neural networks to accurately predict potential energy surfaces and other quantum-chemical properties of molecules and materials, leveraging its efficient handling of large datasets and integration with the Atomic Simulation Environment. (NA?)

  • Use unsupervised and supervised machine learning techniques to analyze large genomics datasets in order to accurately classify triple-negative breast cancer (TNBC) into three distinct subtypes based on your immunogenic profiles, which can then inform personalized treatment decisions. (NA?)

  • Utilise pseudotemporal reordering techniques like Monocle to analyse differentiation from time series data, thereby mitigating the impact of asynchronicity or Simpsons Paradox.’ (NA?)

  • Utilise sparse canonical correlation analysis (sCCA) to identify brain-based dimensions of psychopathology using resting-state functional magnetic resonance imaging (rs-fMRI) data. (NA?)

  • Use crude estimations (CEPs) as additional descriptors alongside traditional chemical descriptors in machine learning (ML) models to improve predictive capabilities, especially when dealing with small datasets. (NA?)

  • Use deep learning algorithms to analyze unlabeled amino acid sequences in order to create a unified representation (UniRep) that can be used to make accurate predictions about protein stability and functionality. (NA?)

  • Take into account the subunit architecture of both ligands and receptors when studying cell-cell communication, as this represents heteromeric complexes accurately and provides a more comprehensive understanding of the underlying biological processes. (NA?)

  • Pay careful attention to the quality of the data used for training and the efficiency of the underlying algorithm when developing machine learning models for biocatalyst design. (NA?)

  • Consider using general-purpose support vector machines (SVM) for accurately classifying cells in single-cell RNA sequencing (scRNA-seq) experiments, as it consistently outperforms other methods in various scenarios. (NA?)

  • Utilize a combination of unbiased feature selection from a reduced-dimension space and machine-learning probability-based prediction method to achieve highly accurate classification of single cells using scRNA-seq data. (NA?)

  • Utilise machine learning methods to effectively integrate diverse sources of information in order to gain a comprehensive understanding of biology and medicine. (NA?)

  • Consider employing advanced computational methods, such as machine learning algorithms, to improve the solubility and membrane permeability of peptide-based therapies, thereby expanding your potential therapeutic applications. (NA?)

  • Carefully consider the choice of benchmarking sets, consensus methods, fragment-based approaches, and machine learning algorithms when conducting molecular docking studies for drug design. (NA?)

  • Adopt a multi-faceted approach combining density functional theory (DFT), high-throughput (HT) computations, and machine learning (ML) techniques to efficiently explore the vast landscape of materials science and identify promising candidates for future technological applications. (NA?)

  • Consider utilizing on-the-fly machine learning schemes for generating force fields automatically during molecular dynamics simulations. This approach can open up the required time and length scales, maintain the chemical precision of first principles methods, and reduce the need for human intervention. (NA?)

  • Consider the potential impact of low complexity sequences on the performance of local alignment tools like BLAST, and explore alternative methods to mitigate this issue, such as masking low complexity regions using applications like SEG. (NA?)

  • Use the single-sample Kullback-Leibler divergence (sKLD) method to effectively detect early warning signals of disease deterioration using high-throughput omics data, even when only a single sample is available. (NA?)

  • Develop interpretable deep learning models like DrugCell, which integrate tumor genotypes and drug structures to predict drug response and uncover biological mechanisms underlying the drug response, thereby improving the accuracy of drug response predictions and facilitating the development of synergistic drug combinations. (NA?)

  • Utilize the DScribe software package for machine learning in materials science, as it offers a comprehensive suite of descriptors that enable accurate and efficient property prediction for atomistic systems. (NA?)

  • Carefully choose and validate your batch-correction algorithm depending on the specific characteristics of your dataset, as different methods perform differently across various scenarios. (NA?)

  • Carefully evaluate the quality of mutation data in publications, resolve ambiguities through author communication whenever necessary, and ensure proper indexing of journals in databases like Medline to improve accessibility and reliability of mutation data. (NA?)

  • Calculate both bootstrap values and concordance factors for the branches on your trees, as these two measures provide complementary information that may help to improve the accuracy of your interpretations of phylogenetic reconstructions. (NA?)

  • Focus on developing mechanism-driven studies of human inherited disease, which have the potential to significantly accelerate the discovery of clinically actionable variants. (NA?)

  • Consider adopting a novel efficient and robust feature selection method that employs joint (_{2,1})-norm minimization on both loss function and regularization, which helps to remove outliers and select features across all data points with joint sparsity. (NA?)

  • Consider using a targeted methylation-based multi-cancer early detection test as a complement to existing single-cancer screening tests, given its high specificity, accurate cancer signal origin prediction, and ability to detect cancer signals across a diverse range of cancers. (NA?)

  • Carefully select appropriate data sources, apply rigorous data cleaning procedures, choose suitable data representation methods, justify model choices with baseline comparisons, and validate models using multiple strategies to avoid overfitting and ensure reliability across diverse applications. (NA?)

  • Utilise Graph Convolutional Networks (GCNs) for protein function prediction, as they offer superior performance compared to other methods, while scaling efficiently to handle large datasets. (NA?)

  • Consider utilizing fourth-generation High Dimensional Neural Network Potentials (4G-HDNNPs) when studying complex systems where long-range charge transfer and multiple charge states play significant roles, as these models offer superior accuracy and flexibility compared to earlier generations of ML potentials. (NA?)

  • Use a machine learning classifier algorithm based on array-generated DNA methylation data to improve the classification of soft tissue and bone tumors, reducing inter-observer variability and misclassification rates. (NA?)

  • Use epigenomic enrichment patterns to understand the underlying tissue-specific mechanisms of complex diseases, such as coronary artery disease (CAD), by partitioning multifactorial trait SNPs into tissue-specific components and examining functional and disease enrichments for distinct subsets of enhancer-overlapping SNPs in each enriched tissue. (NA?)

  • Consider utilizing unsupervised learning techniques when working with large protein sequence datasets, as doing so can lead to significant improvements in representation learning and enable accurate predictions of mutational effects, secondary structure, and long-range contact prediction. (NA?)

  • Consider utilizing a combination of graph neural networks and pseudo bi-level optimization schemes when attempting to integrate both sequential and structural information in protein self-supervised learning. (NA?)

  • Utilise machine learning techniques like support vector machines (SVMs) for improved prediction of MHC class I binding peptides, as they demonstrate superior specificity and broader applicability across various MHC types compared to traditional profile-based methods. (NA?)

  • Use hybrid chemical language models (CLMs) for de novo drug design, incorporating both molecular structure and bioactivity information, to improve the accuracy and efficiency of virtual compound screening and activity-focused molecular design. (NA?)

  • Consider employing quantitative transcriptional co-expression for inference of gene function instead of relying solely on tissue-specific transcription, as it provides a more comprehensive understanding of gene interactions and functionality. (NA?)

Sequence Alignment

  • Utilise the DEDAL algorithm for pairwise sequence alignments, as it offers improved accuracy and efficiency due to its flexible parameterisation and automatic learning capabilities. (Llinares-López et al. 2021)

  • Consider using the RecordLinkage package when working with data from multiple sources, as it offers various techniques for detecting and correcting homonym and synonym errors, thus improving data accuracy and reliability. (Sariyar and Borg 2010)

  • Utilise gradient-based boosting to simplify the complex reward model selection problem inherent in relational sequence alignment, thereby improving the efficiency and effectiveness of the alignment process. (Karwath, Kersting, and Landwehr 2008)

  • Employ a gradient boosting approach to reduce sequence learning to a series of standard function approximation problems, allowing for efficient induction of complex features without reliance on predefined ones. (Dietterich, Ashenfelter, and Bulatov 2004)

  • Consider using the Bayesian probabilistic framework when analyzing complex biological datasets, as it provides a unified approach to various machine learning algorithms and allows for better integration of prior knowledge. (NA?)

  • Consider utilizing a novel approach called “profile-based protein representation” to extract evolutionary information through frequency profiles, which can significantly enhance the performance of sequence-based kernels in protein remote homology detection tasks. (NA?)

Phylogenetic Analysis

  • Utilize the BEAST 2 software platform for Bayesian evolutionary analysis due to its improved flexibility, extensibility, and efficiency compared to previous iterations. (Bouckaert et al. 2014)

  • Carefully consider the choice of support measures in phylogenetic analyses, as both jackknife and Bayesian methods have limitations and potential biases, and neither is universally superior across all scenarios. (Simmons, Pickett, and Miya 2004)

  • Utilize Bayesian inference to effectively handle both mapping and phylogenetic uncertainty in evolutionary reconstruction, thereby improving the accuracy and reliability of your results. (RONQUIST 2004)

  • Integrate over uncertainty in the tree, branch lengths, and substitution model parameters when inferring ancestral states using hierarchical Bayesian methods, rather than relying solely on empirical Bayesian estimates based on fixed assumptions. (Huelsenbeck and Bollback 2001)

  • Move beyond pairwise significance tests when comparing microbial communities, utilizing advanced phylogenetic techniques like UniFrac to understand the underlying causes of community differences. (NA?)

  • Carefully consider the choice of phylogenetic method based on factors like computational efficiency, accuracy, and ability to handle complex evolutionary scenarios, while being aware of potential pitfalls such as long-branch attraction and model misspecification. (NA?)

Protein Structure Prediction

  • Use a combination of deep learning techniques, such as AlphaFold2, along with traditional bioinformatics tools, like CATH-Assign, to accurately identify and classify protein structures, leading to significant advancements in understanding protein function and evolution. (Bordin et al. 2023)

  • Consider utilizing the advanced capabilities of the latest version of AlphaFold, which offers significant improvements in accuracy and expanded functionality for predicting the structures of complexes involving proteins, nucleic acids, small molecules, ions, and modified residues. (Ke Chen et al. 2023)

  • Consider utilising language models trained on protein sequences to achieve rapid and accurate atomic resolution structure prediction, thereby reducing the reliance on multiple sequence alignments and templates. (Zeming Lin et al. 2022)

  • Use deep learning techniques, specifically deep neural networks, to accurately predict protein structures based on sequence data. This approach allows for improved understanding of protein interactions and potentially leads to more effective drug development. (Sheng Wang et al. 2016)

  • Consider utilizing Deep Convolutional Neural Fields (DeepCNF) for protein secondary structure prediction, as it demonstrates superior performance compared to current state-of-the-art methods, particularly in accurately predicting challenging structural elements such as high curvature regions, beta loops, and irregular loops. (Sheng Wang et al. 2015)

  • Carefully select a diverse range of non-homologous proteins with high-resolution structures and α/α domain types to ensure robustness and generalizability of results in studies investigating protein secondary structure. (NA?)

  • Utilize the Support Vector Machine (SVM) method for predicting protein structural classes because it demonstrates high rates of self-consistency and jackknife test, suggesting a strong correlation between protein structural class and amino acid composition. (NA?)

  • Utilize normal mode analysis within a database framework to effectively predict and categorize protein motions based on mode concentration, a novel statistic related to information content. (NA?)

  • Utilise multivariate pattern classification methods to detect subtle and spatially complex patterns of morphological group differences in brain images, which are often undetectable by voxel-based morphometric methods. (NA?)

  • Use a diverse and comprehensive training set when developing algorithms for identifying coiled coils in protein structures, as it leads to improved recognition across a wider range of sequences. (NA?)

  • Consider utilizing a machine learning information retrieval approach to fold recognition, specifically through the application of Support Vector Machines (SVMs) trained on various similarity features extracted from query-template pairs, in order to effectively identify and rank relevant protein templates for accurate structure prediction. (NA?)

  • Consider multiple computational algorithms when attempting to predict intrinsically disordered regions from amino acid sequences, as each algorithm has its own strengths and limitations. (NA?)

  • Utilise image warping first’ workflows and consensus spot patterns in order to improve accuracy and efficiency in 2-D gel image analysis.’ (NA?)

  • Utilize the novel method AntiBP to accurately predict whether a given peptide is antibacterial or not, thereby saving time and resources compared to traditional experimental methods. (NA?)

  • Use a support vector machine to accurately predict protein stability free energy change (()G) upon single point mutation by discriminating between stabilizing, destabilizing, and neutral mutations, achieving an overall accuracy of 56% when performed starting from sequence information and 61% when the protein structure is available. (NA?)

  • Consider utilizing both template-based and sequence-based contact predictions in combination to enhance the accuracy and coverage of protein structure predictions, particularly for template-free modeling targets. (NA?)

  • Consider combining both sequence-based and structure-based methods for protein chemical shift prediction, as this hybrid approach can lead to increased accuracy, broader coverage, and faster calculations compared to either method alone. (NA?)

  • Focus on developing solid-state H+-FET devices that enable electrostatic control over protonic current, utilizing maleic-chitosan nanofiber proton channels bridged by PdHj contacts on a SiO2 gate dielectric, allowing for effective interfacing with biological proton-conducting channels. (NA?)

  • Consider using machine learning techniques, specifically Support Vector Machines (SVMs), to predict substrate specificity of Non-ribosomal Peptide Synthetases (NRPS) Adenylation (A-) domains, achieving high accuracy across multiple hierarchical levels. (NA?)

  • Consider using the Sanjeevini web-server for target-directed lead molecule discovery, as it offers a comprehensive suite of tools for automated detection of active sites, scanning against a vast compound library, all-atom based docking and scoring, and various other utilities to design molecules with desired affinity and specificity against biomolecular targets. (NA?)

  • Utilize the DeepView/Swiss-PdbViewer software to effectively define and search for structural motifs in large protein structure databases, enabling the identification of recurring arrangements of residues with potential structural implications. (NA?)

  • Utilize global statistical approaches to analyze protein sequences, as opposed to local statistical models, in order to effectively remove transitive correlations and accurately predict protein structures. (NA?)

  • Utilise a combination of discovery-driven (shotgun) and hypothesis-driven (targeted) proteomic measurements to achieve near-complete coverage of the yeast proteome, thereby providing a more accurate understanding of protein expression patterns and relationships. (NA?)

  • Utilize the Isotopologue Parameter Optimization (IPO) R-package to enhance the accuracy and efficiency of metabolomics data analysis through optimizing various parameters related to peak picking, retention time correction, and grouping. (NA?)

  • Consider employing an iterative deep learning approach to improve the prediction of multiple structural properties simultaneously, such as secondary structure, torsion angles, Cα-atom based angles and dihedral angles, and solvent accessible surface area. (NA?)

  • Utilise computational methods in drug discovery, particularly computer-aided drug discovery (CADD) tools, to expedite the traditionally lengthy, expensive, and challenging process of drug discovery and development. (NA?)

  • Utilize the Isoelectric Point Calculator (IPC) for accurately estimating the isoelectric point (pl) of proteins and peptides, as it outperforms existing algorithms by at least 14.9% for proteins and 0.9% for peptides, leading to fewer outlier predictions. (NA?)

  • Aim to create a dataset and fitting protocol that is as generic as possible while still achieving your goals, and validate your machine-learned models against reference DFT methods instead of experiments. (NA?)

  • Consider adopting the SeqVec method for analyzing protein sequences, as it offers significant improvements in speed and accuracy over traditional methods like HHblits and Word2vec-like approaches, particularly for tasks such as secondary structure prediction, localization prediction, and distinguishing between membrane-bound and water-soluble proteins. (NA?)

  • Develop physics-inspired structural representations for molecules and materials that capture the inherent symmetries, smoothness, locality, and additivity of atomic arrangements, while maintaining computational efficiency and interpretability. (NA?)

  • Utilise the AlphaFold algorithm for predicting protein structures due to its superior performance in accurately predicting full chains, inter-domain packing, and providing a well-calibrated confidence measure. (NA?)

  • Utilise general protein language models to efficiently identify and test potential affinity-enhancing substitutions in antibodies, leading to successful improvements in binding affinities for all clinically relevant antibodies tested. (NA?)

Gene Expression Analysis

  • Avoid relying solely on conservation filtering when predicting microRNA target sites, as doing so may lead to a substantial loss of bona fide targets. (NA?)

  • Use a state-space model (SSM) when studying complex gene regulatory networks, as it reduces the number of unknown free parameters and minimizes the risk of over-fitting the observed data. (NA?)

Genome Assembly

  • Carefully evaluate the choice of prior distribution and computational approach when implementing Bayesian haplotype reconstruction methods, as these choices significantly impact the accuracy and efficiency of the estimation process. (NA?)

  • Utilise machine learning techniques such as Support Vector Machines (SVMs) and Gaussian Mixture Models to improve the accuracy of base calling in genomic sequencing, particularly for the Illumina Genome Analyser. (NA?)

  • Use a bidirectional recurrent neural network (RNN) with long short-term memory (LSTM) to accurately detect DNA modifications in long-read sequencing data, improving upon traditional methods like bisulfite sequencing and PacBio long-read sequencing. (NA?)

Metagenomic Data Analysis

  • Utilise model-based methods, specifically hierarchical Bayesian models, for analyzing multivariate abundance data in community ecology, as they allow for a data-generating process and likelihood function that can be tailored to match specific ecological processes and questions of interest. (Hui 2016)

  • Utilise the Ballgown suite as a bridge between upstream assembly tools and downstream statistical modeling tools in Bioconductor. (Frazee et al. 2014)

  • Use the Viral Informatics Resource for Metagenome Exploration (VIROME) pipeline to comprehensively classify and characterize viral metagenome sequences based on homology search results against both known and environmental sequences, thereby enabling accurate functional and taxonomic information derivation from various annotated sequence databases. (NA?)

  • Carefully consider and optimize quality-filtering parameters when analyzing Illumina amplicon sequencing data to improve accuracy in estimating microbial diversity. (NA?)

  • Use Calypso, a user-friendly web-based platform, to analyze and visualize complex microbiome-environment interactions through various multivariate statistical techniques, enabling comprehensive understanding of these intricate relationships. (NA?)

  • Consider combining microbiota analysis with existing screening methods, such as the fecal immunochemical test (FIT), to improve the detection of colonic lesions, particularly adenomas. (NA?)

  • Consider using a k-mer based tool like VirFinder for identifying prokaryotic viral sequences from metagenomic data, particularly for short contigs, as it shows improved performance compared to VirSorter in correctly identifying novel viruses. (NA?)

  • Avoid assigning sequences to novel groups that are absent from the training set, as this leads to high over classification rates, especially in microbiome studies where many microorganisms are not represented in reference taxonomies. (NA?)

  • Carefully consider and address potential sources of bias such as batch effects and contamination in your experimental designs, particularly when working with large datasets like TCGA, to ensure accurate and reliable results. (NA?)

  • Carefully consider the advantages and limitations of various high-throughput sequencing (HTS) methods for microbiome analysis, such as amplicon sequencing, metagenomic sequencing, and metatranscriptomic sequencing, when choosing the appropriate approach for your specific research question and sample type. (NA?)

  • Consider employing deep learning techniques, specifically convolutional neural networks, for accurate identification of viral sequences in metagenomic data, as demonstrated by the superior performance of DeepVirFinder compared to traditional methods. (NA?)

Computational Chemistry

  • Focus on developing a generative model for predicting reaction mechanisms, specifically for linear electron flow (LEF) reactions, which enables accurate predictions of reaction outcomes while providing valuable insights into the underlying chemical processes. (Schwaller et al. 2018)

  • Use MoleculeNet, a large-scale benchmark for molecular machine learning, to evaluate the efficacy of proposed methods, as it enables comparisons across different datasets and algorithms, thereby improving the overall quality of research in the field. (Zhenqin Wu et al. 2017)

  • Carefully choose both the regressor and molecular representation when constructing fast machine learning (ML) models for predicting electronic ground-state properties of organic molecules, as the out-of-sample errors are highly dependent on these choices. (Faber et al. 2017)

  • Carefully choose the appropriate threshold for defining active’ compounds in your studies, as setting too low a threshold can result in a skewed distribution of ‘active’ and ‘inactive’ compounds, potentially affecting the reliability and generalizability of findings.’ (Lenselink et al. 2017)

  • Combine heuristic and chemical rules to reduce the number of theoretical formulas to a small set of the most likely compositions, while considering factors such as element numbers, Lewis and Senior checks, isotopic pattern filters, and other relevant constraints. (NA?)

  • Consider using the newly introduced Graph Fragment (GF)-based descriptors for chemical compound classification and retrieval, as they provide a more comprehensive and accurate representation of the underlying molecular structures compared to traditional fingerprint-based descriptors. (NA?)

  • Consider utilising machine learning algorithms to predict atomisation energies of organic molecules based solely on nuclear charges and atomic positions, offering significant time savings and improved accuracy compared to traditional methods. (NA?)

  • Consider using ligand efficiency metrics like ligand efficiency (LipE), ligand lipophilic efficiency (LLE), and logP divided by ligand efficiency (LIEP) when optimising lead compounds, as they take into account both size and lipophilicity, helping to identify compounds that make efficient use of your chemical structure for desired binding, leading to improved drug candidates. (NA?)

  • Utilize the Competitive Fragmentation Modeling (CFM) technique for accurate and efficient identification of metabolites in electrospray tandem mass spectrometry (ESI-MS/MS) data. (NA?)

  • Consider utilizing the ab initio nanoreactor, a highly accelerated, first-principles molecular dynamics simulation, to discover new molecules and mechanisms without predetermining reaction coordinates or elementary steps, thereby expanding the scope of theoretical and computational chemistry beyond merely interpreting experimental findings. (NA?)

  • Carefully evaluate the transferability of computational models across diverse chemical spaces, considering factors like molecular size, chemical composition, and atomistic configurations, while also utilizing experimental or high-level quantum chemistry results for validation purposes. (NA?)

  • Utilise the Spectral Neighbourhood Analysis Potential (SNAP) method when developing interatomic potentials for solids and liquids, as it employs machine-learning techniques to reproduce the energies, forces, and stress tensors of a large set of small configurations of atoms, resulting in a more accurate representation of the behaviour of these materials. (NA?)

  • Utilize multiple kernel learning and support vector machines to effectively identify unknown compounds through your MS/MS spectra and fragmentation trees, significantly improving upon existing methods. (NA?)

  • Utilize smart automation throughout the entire molecular design cycle, combining machine learning methods with miniaturization and lab-on-a-chip technology, to enhance the efficiency and effectiveness of drug discovery. (NA?)

  • Focus on developing a systematic set of descriptors for compounds that satisfy certain criteria, such as being applicable to a wide range of chemical compositions and crystal structures, while ensuring translational and rotational invariance. (NA?)

  • Use a large benchmark dataset containing multiple assays to ensure accurate evaluation of machine learning methods for compound target prediction, while addressing issues like compound series bias and hyperparameter selection bias through techniques like cluster-cross-validation. (NA?)

  • Consider using MoleculeNet, a benchmark collection for molecular machine learning, to standardize your approach to developing and improving models for learning molecular properties. (NA?)

  • Carefully choose your molecular descriptor calculation software based on factors like the number of descriptors supported, ease of installation and use, calculation speed, availability of automated tests, and compatibility with various programming languages and environments. (NA?)

  • Consider incorporating physical symmetries (spatial, temporal, and local symmetries) into a gradient-domain machine learning approach to significantly improve the accuracy and efficiency of molecular simulations. (NA?)

  • Employ active learning strategies in order to optimize the efficiency and effectiveness of your machine learning models. (NA?)

  • Employ advanced AI-based sophisticated machine learning tools to analyze and systemize larger data sets in order to improve the efficiency and cost-effectiveness of structure-based drug discovery. (NA?)

  • Consider utilizing BioTransformer, a comprehensive computational tool combining machine learning and knowledge-based approaches, for accurate, rapid, and extensive in silico metabolism prediction and compound identification across multiple domains, including human tissues, gut microbiota, and the environment. (NA?)

  • Consider implementing an on-the-fly machine learning force field generation technique for simulating phase transitions, as it significantly reduces the computational cost while maintaining accuracy. (NA?)

  • Utilise machine-learning interatomic potentials with active learning to accelerate crystal structure prediction, allowing for an automated construction of an interatomic interaction model from scratch, thereby reducing the need for expensive density functional theory calculations. (NA?)

  • Utilize a comprehensive chemoinformatics-guided workflow to identify and optimize selective catalysts, incorporating advanced machine learning techniques and robust molecular descriptors to ensure accurate predictions even in situations where the target selectivity lies beyond the bounds of the training data. (NA?)

  • Consider using the open-source AiZynthFinder software for retrosynthesis planning, as it offers a transparent, flexible, and efficient approach to generating synthetic routes for various compounds. (NA?)

  • Utilize the SELFIES (Self-referencing Embedded Strings) molecular representation instead of SMILES (Simplified Molecular Input Line Entry System) for generating valid molecules in machine learning models, as SELFIES guarantees 100% validity and greater diversity compared to SMILES. (NA?)

  • Combine computational chemistry (CompChem) and machine learning (ML) techniques effectively to achieve transformative impacts on chemical sciences, particularly through improving computational algorithms and amplifying insights available from computational chemistry methods. (NA?)

Optimization Techniques

  • Carefully balance the tradeoff between the quality of approximation and computational efficiency when implementing quasi-Newton techniques in stochastic gradient descent algorithms. (Bietti et al. 2023)

  • Consider using evolutionary search with warm-starts and restarts, combined with abstract execution for pruning redundancy, to efficiently navigate the vast and sparse program search space when attempting to discover novel optimization algorithms. (Xiangning Chen et al. 2023)

  • Consider using the D-Adaptation technique to optimize your algorithms, particularly in cases where the optimal learning rate is unknown or difficult to determine. (Defazio and Mishchenko 2023)

  • Consider using metaheuristics as a promising approach to optimize and customize large language models through prompt learning, as it satisfies the criteria of being automatic, discrete, black-box, gradient-free, and interpretable. (R. Pan et al. 2023)

  • Utilise a learning rate proportional to the inverse of the number of iterations, combined with averaging, to achieve optimal convergence rates in stochastic approximation problems involving machine learning algorithms such as kernel logistic regression and least-squares regression. (Godichon-Baggioni, Werge, and Wintenberger 2023)

  • Consider using the No Free Lunch (NFL) theorem when analyzing various domains like search, bandits, self-play, coevolution, generalized optimization, and supervised learning, as it provides insights into the limits of performance guarantees across different algorithms and highlights the importance of understanding the underlying assumptions and constraints within these domains. (D. H. Wolpert 2023)

  • Carefully consider the geometry of your data and apply appropriate methods to analyze high-dimensional convex structures, taking into account concentration phenomena and the properties of various polytopes. (“Proceedings of Third International Conference on Sustainable Expert Systems” 2023)

  • Consider utilizing automated machine learning (AutoML) techniques to efficiently and effectively manage the vast quantities of IoT data being generated, particularly in dynamic environments where concept drift may occur. (Li Yang and Shami 2022)

  • Carefully examine the relationship between the structure allowing for the avoidance of barren plateaus and the potential for efficient classical simulation, as the former may also facilitate the latter, raising doubts about the non-classicality of information processing capabilities in parametrized quantum circuits and the possibility of achieving superpolynomial advantages from running them on quantum hardware. (Elben et al. 2022)

  • Consider the unique challenges presented by TinyML systems, such as low power consumption, limited memory, hardware heterogeneity, software heterogeneity, and cross-product compatibility, when developing a comprehensive and accurate benchmarking suite for evaluating the performance of these systems. (Banbury et al. 2021)

  • Ensure pathway uniqueness for the stochastic differential equation before attempting to construct a strong solution using methods like Eulers approximations. (Gyöngy and Krylov 2021)

  • Consider using deep learning compilers that employ advanced optimization techniques, such as weight pruning, quantization schemes, and math kernel libraries, to improve the performance and efficiency of your deep learning models on embedded platforms. (Sponner, Waschneck, and Kumar 2021)

  • Utilise a black-box prompt tuning framework for vision-language models (VLMs) in order to learn task-relevant prompts without back-propagation. (Dosovitskiy et al. 2020)

  • Carefully consider the various levels of automation in AutoML systems when developing and evaluating them, as well as the interactions among the optimizer, meta-learner, and data-model processing methods, in order to effectively address the challenges and opportunities in this growing field. (Escalante 2020)

  • Focus on developing flexible models that incorporate both causal inference and robustness, utilizing computational advancements to efficiently analyze complex data sets. (Andrew Gelman and Vehtari 2020)

  • Use a transformer-structured configuration searcher enhanced with multi-head attention and memory mechanism to efficiently locate high-performance configurations in a vast search space, thereby improving hyper-parameter optimization for deep neural networks. (Yimin Huang et al. 2020)

  • Focus on identifying well-performing general-purpose optimizers for deep learning, especially when there is no prior knowledge about well-working hyperparameter values for each specific problem. (R. M. Schmidt, Schneider, and Hennig 2020)

  • Consider integrating GPU-accelerated computation into your workflows, specifically through the use of the OpenCL framework, to achieve significant speedups in complex mathematical tasks such as those encountered in Bayesian modelling and inference. (Češnovar et al. 2019)

  • Prioritise the selection of important gradients over redundant ones during the training process of large-scale deep learning models on ring structures, thereby reducing bandwidth usage while preserving training accuracy. (Zehua Cheng and Xu 2019)

  • Consider framing adversarial training as a discrete time differential game, allowing them to analyze the Pontryagins Maximum Principle (PMP) of the problem, which reveals that the adversary update is only coupled with the parameters of the first layer of the network. This insight inspired the authors to develop the YOPO (You Only Propagate Once) algorithm, which restricts most of the forward and back propagation within the first layer of the network during adversary (Dinghuai Zhang et al. 2019)

  • Consider using the SYCL programming model for developing machine learning applications on OpenCL hardware due to its ability to adapt to various hardware characteristics and optimize performance across multiple devices. (R. Burns et al. 2019)

  • Consider formulating adversarial training problems as differential games and deriving the Pontryagins Maximum Principle (PMP) to optimize the adversarial perturbation, leading to reduced computational costs and improved efficiency.’ (Askari et al. 2018)

  • Utilize machine learning techniques to develop algorithms that can effectively solve combinatorial optimization problems by learning from a chosen implicit distribution of problems, thereby increasing efficiency and reducing computational costs. (Yoshua Bengio, Lodi, and Prouvost 2018)

  • Carefully choose informative, interpretable, cheaply computable, generally applicable, and complementary features for your datasets to improve the effectiveness of automated algorithm selection systems. (Kerschke et al. 2018)

  • Use the reliable fraction of information (RFI) metric instead of the traditional fraction of information (FOI) metric for discovering dependencies in high-dimensional datasets because RFI corrects for the “dependency-by-chance” bias inherent in FOI estimation, leading to more accurate and robust results. (Mandros, Boley, and Vreeken 2018)

  • Consider using sparsified SGD with memory to improve the efficiency of your machine learning algorithms while maintaining the same convergence rate as traditional SGD. (Stich, Cordonnier, and Jaggi 2018)

  • Consider using the StructADMM framework for structured weight pruning in deep neural networks, as it enables various types of structured sparsity, guarantees solution feasibility, provides high solution quality, and significantly improves weight pruning rate while preserving accuracy. (Tianyun Zhang, Ye, Zhang, Ma, et al. 2018)

  • Utilize a novel optimization method called Diffused Stochastic Gradient Descent (D-SGD) for efficiently handling highly-dynamic and recency-sensitive data, which involves assigning recency-sensitive weights to different samples and selecting samples accordingly for gradient calculations, followed by updating related samples via a diffusion strategy. (Xumin Chen et al. 2018)

  • Consider using mixed HMC (M-HMC) as a general framework to efficiently sample from complex distributions with mixed discrete and continuous variables, enabling more frequent updates of discrete variables while retaining HMCs ability to suppress random-walk behavior.’ (Betancourt 2017)

  • Use a multi-stage procedure called “Coarse-ID control” to optimize control cost in situations where the dynamics of the system being controlled are unknown. This involves estimating a model from a limited number of experimental trials, assessing the accuracy of that model against the actual system, and then designing a controller that takes into account both the model and its associated uncertainty. (Dean et al. 2017)

  • Utilise Stochastic Gradient Descent (SGD) as an approximate Bayesian posterior inference algorithm, specifically by adjusting the tuning parameters of constant SGD to best match the stationary distribution to a posterior, thereby minimising the Kullback-Leibler divergence between these two distributions. (Dieuleveut, Durmus, and Bach 2017)

  • Carefully choose between parametric and nonparametric methods for frontier estimation, considering factors like the need for assumptions about functional form, the ability to handle multiple inputs and outputs, and the potential impact on results. (Podinovski 2017)

  • Incorporate randomization techniques during the inference stage to enhance the robustness of your Convolutional Neural Networks (CNNs) against adversarial attacks. (C. Xie et al. 2017)

  • Utilise the Reluplex algorithm, which extends the simplex algorithm to handle the non-convex Rectified Linear Unit (ReLU) activation function, in order to efficiently verify properties of deep neural networks (DNNs) without making simplifying assumptions. (G. Katz et al. 2017)

  • Carefully consider the regularity assumptions of your data when developing algorithms for distributed optimization, as different assumptions can significantly affect the performance and convergence rates of the resulting algorithms. (G. Lan, Lee, and Zhou 2017)

  • Focus on developing algorithms that leverage global smoothness assumptions to achieve faster rates of convergence on globally smooth problems, rather than relying solely on local smoothness assumptions. (Malherbe and Vayatis 2017)

  • Carefully consider the trade-off between computation and communication costs when selecting the appropriate distributed optimization algorithm and cluster size for a given problem. (X. Pan et al. 2017)

  • Consider using a perturbed form of gradient descent to efficiently escape saddle points in non-convex optimization problems, achieving near “dimension-free” convergence to second-order stationary points. (N. Agarwal et al. 2016)

  • Consider employing a novel convexification-decomposition technique combined with dynamic consensus to address the challenge of solving nonconvex distributed optimization problems in multi-agent networks with time-varying connectivity. (Lorenzo and Scutari 2016)

  • Utilise CVXPY, a powerful tool for convex optimisation, which offers improvements like signed DCP for verifying convexity, parameters for handling constant values, and an object-oriented approach for constructing complex optimisation problems. (Diamond and Boyd 2016)

  • Carefully evaluate the generalization capabilities of adaptive gradient methods compared to traditional gradient descent or stochastic gradient descent methods before choosing an optimization approach for training deep neural networks. (Isola et al. 2016)

  • Focus on improving the quality of the output signal while reducing the amount of hardware required for filtering, particularly when dealing with applications such as embedded FIR filters in medical devices. (Meidani and Mashoufi 2016)

  • Use recursive decomposition to tackle complex nonconvex optimization problems, as it offers significant improvements in efficiency compared to traditional approaches. (Friesen and Domingos 2016)

  • Utilize feedback from the objective function to dynamically adjust the learning rate in stochastic gradient descent algorithms, leading to improved optimization performance. (Hayashi, Koushik, and Neubig 2016)

  • Develop an end-to-end compiler, TVM, capable of taking high-level specifications of deep learning programs from existing frameworks and generating low-level optimized code for a diverse set of hardware back-ends, thereby offering performance comparable to manually optimized operator libraries across various hardware back-ends. (K. He et al. 2016)

  • Utilise the Hyperband algorithm for hyperparameter optimization, as it speeds up random search through adaptive resource allocation and early-stopping, offering over an order-of-magnitude speedup over popular Bayesian optimization methods on a variety of deep-learning and kernel-based learning problems. (Lisha Li et al. 2016)

  • Utilize the ZipML framework to enable end-to-end low precision training of machine learning models, leading to significant reductions in computation and communication costs. (Hantian Zhang et al. 2016)

  • Utilize incremental methods for minimizing sums of convex component functions, as these methods have proven highly effective in practice and can be adapted to a wide range of application areas through a unified algorithmic framework. (Bertsekas 2015)

  • Pay close attention to the challenges associated with defining, validating, and updating machine learning models, particularly in terms of managing feature transformations, implicit assumptions, and potential adversarial settings. (Tianqi Chen et al. 2015)

  • Utilise Bayesian neural networks for Bayesian optimization, enabling scalability and robustness through stochastic gradient Hamiltonian Monte Carlo, improved further by a scale adaptation technique. (Y. Gal and Ghahramani 2015)

  • Consider combining online Monte Carlo methods with model distillation to achieve a simple, scalable approach to Bayesian inference of the parameters of neural networks, leading to improved log likelihood scores on the test set compared to traditional methods like Stochastic Gradient Descent (SGD) and Expectation Propagation (EP). (Korattikara et al. 2015)

  • Consider replacing traditional Gaussian processes with deep neural networks in Bayesian optimization to achieve better scalability and efficiency, allowing for faster and more accurate optimization of complex machine learning models. (Snoek et al. 2015)

  • Carefully evaluate the tradeoff between communication costs and computational complexity when choosing between centralized and decentralized algorithms for distributed machine learning tasks, particularly in situations with limited network resources. (Aybat et al. 2015)

  • Consider utilising Autograd, a package that enables automatic differentiation within standard Python and NumPy code, thereby significantly reducing the time spent on writing gradients and inference procedures, thus accelerating the overall research process. (Baydin et al. 2015)

  • Utilise the equilibration preconditioner rather than the Jacobi preconditioner when dealing with non-convex optimization problems, as it demonstrates superior performance in reducing the condition number and avoiding divergence caused by underestimation of curvature. (Dauphin, Vries, and Bengio 2015)

  • Leverage the increasing availability of data to create statistically informed uncertainty sets for robust optimization, leading to improved decision making under uncertainty. (Bertsimas, Gupta, and Kallus 2014a)

  • Utilise the Minimization by Incremental Surrogate Optimisation (MISO) scheme for efficient handling of large-scale machine learning problems. (Mairal 2014)

  • Use non-asymptotic moment estimates and concentration inequalities to analyze the rate of convergence of empirical measures to your underlying distributions, particularly when considering the Wasserstein distance of order \(p > 0\). (Fournier and Guillin 2014)

  • Prioritize optimizing memory transfer efficiency alongside computational acceleration when designing machine-learning accelerators, as inefficient memory transfers can negate the benefits of computational acceleration. (Tianshi Chen et al. 2014)

  • Avoid naive clamping and squashing functions when enforcing control limits in differential dynamic programming, and instead adopt a projected-Newton QP solver to efficiently handle box inequality constraints. (Tassa, Mansard, and Todorov 2014)

  • Consider decoupling the feature space into a pair of complementary subspaces - stability space and plasticity space - to achieve a better balance between stability and plasticity in continual learning algorithms. (Bahdanau, Cho, and Bengio 2014)

  • Focus on developing a black box’ variational inference algorithm that enables quick application to multiple models with minimal additional derivation, thereby facilitating rapid development and exploration of diverse models for addressing complex problems.’ (Ranganath, Gerrish, and Blei 2014)

  • Carefully consider the assumptions of variable independence, redundancy in network parametrization, and uniformity when studying the connection between the loss function of a fully-connected feed-forward neural network and the Hamiltonian of the spherical spin-glass model. (Choromanska et al. 2014)

  • Utilize a proximal version of the stochastic dual coordinate ascent method combined with an inner-outer iteration procedure to accelerate the process, resulting in improved runtimes for various machine learning optimization problems. (Shalev-Shwartz and Zhang 2013)

  • Utilize the novel semidifferential framework for submodular function optimization, which combines sub- and super-differentials to provide a unified approach for both submodular minimization and maximization, offering improved efficiency and scalability for machine learning applications. (Iyer, Jegelka, and Bilmes 2013)

  • Consider employing advanced optimization techniques, such as operator splitting approaches like the alternating direction method of multipliers (ADMM), for inverting deep neural networks, as they can potentially lead to improvements in both iteration count and solution quality compared to traditional gradient descent methods. (Szegedy et al. 2013)

  • Focus on optimizing hyper-parameters, utilizing unsupervised representation learning techniques, and employing stochastic gradient descent methods for efficient and effective training of deep neural networks. (Yoshua Bengio 2012)

  • Consider utilizing parallel coordinate descent methods (PCDMs) for optimizing big data problems, particularly when the objective function exhibits partial separability, as this approach can significantly reduce the time required for convergence. (Richtárik and Takáč 2012)

  • Utilize the Stochastic Average Gradient (SAG) method for optimizing the sum of a finite set of smooth, strongly convex functions, as it offers a linear convergence rate while maintaining the low iteration cost of traditional stochastic gradient methods. (Roux, Schmidt, and Bach 2012)

  • Carefully consider the choice of your optimization algorithm, specifically focusing on the Stochastic Gradient Descent (SGD) method, which offers promising results for non-smooth optimization tasks, while also exploring the potential benefits of incorporating a running average scheme to further enhance the optimization accuracy. (Shamir and Zhang 2012)

  • Consider an incremental learning-to-learn approach for improving your models performance on future tasks, rather than solely focusing on traditional batch learning-to-learn techniques. (Hazan and Kale 2012)

  • Carefully combine multiple techniques to create an efficient implementation for learning linear predictors with convex losses on terascale data sets, achieving significant scalability and efficiency improvements. (Alekh Agarwal et al. 2011)

  • Utilise asynchronous gradient methods in distributed optimization scenarios, as these methods can achieve asymptotically optimal rates for stochastic convex optimization despite potential delays caused by asynchronicity. (Alekh Agarwal and Duchi 2011)

  • Utilise submodular functions in machine learning due to your ability to express problems directly, provide useful regularisation functions for supervised and unsupervised learning, and enable the development of efficient algorithms for approximate and exact submodular function minimisation with theoretical guarantees and good practical performance. (F. Bach 2011)

  • Carefully select appropriate Gaussian processes priors and estimate your parameters accurately to ensure optimal convergence rates in efficient global optimization problems. (Bull 2011)

  • Pay close attention to the initialization and momentum parameters when employing Stochastic Gradient Descent with Momentum (SGDM) for training deep and recurrent neural networks, as proper tuning of these parameters can significantly enhance the performance of the model. (Cotter et al. 2011)

  • Consider using the No-U-Turn Sampler (NUTS) algorithm instead of traditional Hamiltonian Monte Carlo (HMC) methods, as NUTS eliminates the need to manually set the number of steps parameter and can improve overall efficiency in sampling from complex distributions. (M. D. Hoffman and Gelman 2011)

  • Prioritize asynchronous algorithms utilizing one-directional (push-based) communications and not rely on doubly-stochastic consensus parameters for creating a robust and efficient implementation of consensus-based algorithms for distributed optimization. (Jakovetic, Xavier, and Moura 2011)

  • Consider implementing Hogwild!, a lock-free approach to parallelizing Stochastic Gradient Descent (SGD), particularly for sparse optimization problems, as it offers nearly optimal rates of convergence and outperforms alternative schemes that use locking by an order of magnitude. (F. Niu et al. 2011)

  • Utilize the Kronecker-factored Approximate Curvature (K-FAC) method when optimizing neural networks, as it provides an efficient approximation of natural gradient descent without sacrificing performance. (Ollivier et al. 2011)

  • Utilise the truncated power method’, a novel approach to solving the sparse eigenvalue problem, which involves applying the classical power method with an added truncation operation to guarantee sparsity. This method offers significant advantages over existing approaches, including improved accuracy and computational efficiency, particularly in cases where the true matrix has a sparse or approximately sparse dominant eigenvector.’ (X.-T. Yuan and Zhang 2011)

  • Utilise the novel concept of discrepancy between functions’ to transform problems of stochastic convex optimization into statistical parameter estimation problems. This allows for more effective use of information-theoretic methods to derive tighter minimax complexity estimates for various function classes.’ (Alekh Agarwal et al. 2010)

  • Consider using the No-U-Turn Sampler (NUTS) algorithm instead of traditional Hamiltonian Monte Carlo (HMC) methods, as it eliminates the need to manually set the number of steps parameter and provides comparable or better efficiency in sampling from high-dimensional target distributions. (Beskos et al. 2010)

  • Use a combination of deterministic and randomized algorithms to achieve optimal convergence rates in stochastic strongly-convex optimization problems, rather than relying solely on online-to-batch conversions. (Hazan and Kale 2010)

  • Utilize the “kernel method” to effectively analyze and understand the behavior of directed lattice paths, providing valuable insights into your enumeration and asymptotics. (Banderier and Nicodème 2010)

  • Utilize a two-step lookahead approach in your Bayesian optimization processes, as it offers improved query efficiency and robustness compared to traditional one-step lookahead methods. (Brochu, Cora, and Freitas 2010)

  • Utilize structured sparsity-inducing norms in order to effectively manage complex data structures and improve model performance through better control of variable selection. (Jenatton, Audibert, and Bach 2009)

  • Utilize the BUGS (Bayesian inference using Gibbs sampling) software for efficient and accurate analysis of complex data sets, as it enables automatic construction and sampling from full conditional distributions, thereby reducing the complexity of handling multiple interrelated unknown parameters and missing data. (Rue, Martino, and Chopin 2009)

  • Consider using the Metropolis-Hastings Robbins-Monro (MH-RM) algorithm for high-dimensional maximum marginal likelihood exploratory item factor analysis, as it provides accurate estimates and has advantages over existing methods such as numerical quadrature based EM algorithm. (L. Cai 2009)

  • Consider adopting a multi-stage convex relaxation scheme for solving problems with non-convex objective functions, particularly in the context of learning formulations with sparse regularization, as it offers improved performance compared to standard L1 convex relaxation. (Tong Zhang 2009)

  • Utilise automated algorithm configuration methods, specifically ParamILS, to efficiently and systematically explore large design spaces and identify optimal parameter configurations for complex algorithms. (Hutter et al. 2009)

  • Consider using the nuclear norm minimization problem (1.7) as a convex and computationally tractable approximation to the matrix rank minimization problem (1.1), particularly when dealing with noisy data or large dimensions. (W. Dai and Milenkovic 2008)

  • Utilise the truncated gradient’ method for inducing sparsity in the weights of online-learning algorithms with convex loss functions. This method offers a continuous degree of sparsity, is theoretically motivated, and works well empirically, making it ideal for handling large datasets with numerous features.’ (Langford, Li, and Zhang 2008)

  • Utilize a combination of object-oriented programming concepts, design patterns, and a framework-based approach to develop a robust, reusable, and extensible software system for evolutionary computation. (S. Ventura et al. 2007)

  • Consider applying Stochastic Meta-Descent (SMD), a stochastic gradient optimization method with gain vector adaptation, to train Conditional Random Fields (CRFs) for faster convergence and higher accuracy. (Vishwanathan et al. 2006)

  • Utilise Gaussian processes and mutual information optimisation algorithms to determine the ideal placement of sensors in a network, ensuring accurate and efficient data collection. (X. Bai et al. 2006)

  • Utilize a combination of local search algorithms and critical clause analysis to efficiently identify minimally unsatisfiable subformulas (MUS) and maximally satisfiable subsets (MSS) within sets of Boolean clauses. (Bruni 2004)

  • Consider utilizing semidefinite programming relaxations for semialgebraic problems, as they offer a hierarchy of convex relaxations that can prove infeasibility for questions reducible to a finite number of polynomial equalities and inequalities. (Parrilo 2003)

  • Utilize path sampling, a novel approach that extends traditional importance sampling techniques by incorporating multiple “bridging” densities, thereby reducing Monte Carlo errors and improving overall accuracy in estimating normalizing constants. (Andrew Gelman and Meng 1998)

  • Ensure your statistical models satisfy the monotonicity condition to guarantee the existence and uniqueness of the solution to forward-backward stochastic differential equations, especially when dealing with cases where the non-degeneracy condition for the forward equation does not apply. (Y. Hu and Peng 1995)

  • Utilize a Four Step Scheme’ to solve forward-backward stochastic differential equations (SDEs) explicitly, ensuring the adaptive solution can be sought in an ‘ordinary’ sense over an arbitrarily prescribed time duration. (J. Ma, Protter, and Yong 1994)

  • Consider using a deep reinforcement learning approach to minimize the execution cost of neural network computation graphs in an optimizing compiler, which involves training an optimizer offline and then generalizing to previously unseen graphs without further training. (Bean 1994)

  • Utilise sample compression schemes to achieve near-optimal sample complexity bounds for learning mixtures of Gaussians. (Blumer et al. 1989)

  • Focus on identifying the optimal stopping time for your experiment based on the minimum expected cost, while considering the constraints imposed by the experimental setup. (Kliemann 1987)

  • Use an optimal linkage rule to minimize the probability of false matches while maximizing the probability of true matches when comparing records across multiple datasets. (Fellegi and Sunter 1969)

  • Utilize machine learning methods such as the proposed IID and Markov models to optimize your iterative searches, thereby improving efficiency and reducing computational costs. (NA?)

  • Consider using hybrid methods combining neural networks and particle swarm optimization techniques to improve problem solving capabilities across various domains. (NA?)

  • Focus on identifying the optimal stopping time for your experiment based on the minimum expected cost, while considering the constraints imposed by the experimental setup. (NA?)

  • Carefully consider the problem encoding and evaluation function when implementing genetic algorithms, as these factors significantly impact the effectiveness of the search and optimization processes. (NA?)

  • Carefully consider the trade-off between the number of structures evaluated and the accuracy of those evaluations in noisy environments, as sometimes less precise evaluations can lead to more efficient searches. (NA?)

  • Consider the potential advantages of using simple random mutation instead of complex genetic operators like crossover and inversion in evolutionary simulations, especially in cases where the goal is to optimize behavior in various environments. (NA?)

  • Carefully evaluate the potential impact of differences between the simulation model used for training and the actual target environment in which the learned decision rules will be implemented. (NA?)

  • Utilize genetic programming techniques to efficiently search through the space of potential computer programs to identify the most suitable candidate for solving a specific problem, leveraging the principles of natural selection and genetic crossover to optimize the search process. (NA?)

  • Carefully consider the problem encoding and evaluation function when implementing genetic algorithms, as these factors significantly impact the efficiency and effectiveness of the optimization process. (NA?)

  • Utilise the concept of Hypergradients’, which enables the computation of gradients with respect to all continuous training parameters, thus enabling efficient storage of necessary information and optimisation of validation loss with respect to thousands of hyperparameters.’ (NA?)

  • Consider using the forward-backward splitting method for efficiently solving regularized convex programming problems, particularly those involving non-differentiable regularization functions like the (_{1})-norm, due to its ability to generate sparse solutions at minimal computational cost and its proven convergence properties. (NA?)

  • Utilize a mobile-agent-based, online Monte Carlo technique inspired by the concept of stigmergy, which involves indirect communication among individuals via environmental modifications, to develop efficient and adaptive routing algorithms for communication networks. (NA?)

  • Consider employing hybrid methods combining both neural networks and particle swarm optimization techniques for improved problem solving capabilities across various domains. (NA?)

  • Consider employing a meta-modelling approach to support automated hyperparameter optimization, aiming to replace hand-tuning with a reproducible and unbiased optimization process. (NA?)

  • Carefully balance the tradeoff between the quality of approximations used in stochastic gradient descent algorithms and the computational efficiency gained through sparse representations, while also considering other speedup opportunities like exploiting sparsity of patterns and improving implementation details. (NA?)

  • Utilise a unifying framework called Model-Based Search’ when dealing with complex combinatorial optimization problems. This framework encompasses various metaheuristics including Ant Colony Optimisation, Stochastic Gradient Ascent, Cross-Entropy Method, and Estimation of Distribution Algorithms, allowing for a more comprehensive understanding and application of these methods.’ (NA?)

  • Consider using optimal feedback control as a theoretical framework for studying the neural basis of voluntary motor control, as it allows for a better integration of motor behavior, limb mechanics, and neural control. (NA?)

  • Consider using the adaptive weighted-sum method instead of the traditional weighted-sum method for bi-objective optimization, as it addresses the limitations of the latter by producing well-distributed solutions, finding Pareto optimal solutions in non-convex regions, and neglecting non-Pareto optimal solutions. (NA?)

  • Carefully consider the type of multi-objective optimization problem they are facing, including the number of objective functions, inequality and equality constraints, and the nature of the design variables, when choosing among various optimization techniques. (NA?)

  • Utilize the cross-entropy (CE) method for solving combinatorial and multi-extremal optimization problems, as well as for rare event simulations, due to its efficiency, simplicity, and ability to provide fast and near-optimal solutions. (NA?)

  • Utilise the cross-entropy method for solving complex continuous multi-extremal optimization problems due to its efficiency, ease of programming, and consistent accuracy compared to other global optimization heuristics. (NA?)

  • Use the Differential Evolution Markov Chain (DE-MC) algorithm for efficient and effective sampling from complex, high-dimensional posteriors in Bayesian statistics. (NA?)

  • Use early stopping rules in your gradient descent algorithms to balance the bias-variance trade-off, leading to improved rates of convergence and better performance in various applications. (NA?)

  • Focus on balancing the interpretability-accuracy tradeoff in the design of fuzzy rule-based classifiers by considering both the number of correctly classified training patterns and the complexity of the model, including the number of fuzzy rules and the total rule length. (NA?)

  • Consider using various mathematical tools and techniques, including Markov chains, tail inequalities, and other randomized algorithm analysis tools, to effectively study the time complexity of evolutionary algorithms for combinatorial optimization problems. (NA?)

  • Carefully choose your optimization techniques for L1-regularized optimization problems depending on the specific characteristics of your dataset and desired outcome, considering factors like computational efficiency, interpretability, and robustness. (NA?)

  • Focus on developing efficient algorithms for solving large-scale structured convex optimization problems, particularly those involving noisy data, while considering the tradeoff between approximation accuracy and computational complexity. (NA?)

  • Consider using the Pegasos algorithm, a simple stochastic sub-gradient descent method, for solving Support Vector Machine (SVM) optimization problems, as it offers fast convergence rates and low computational costs compared to other methods. (NA?)

  • Carefully consider the choice of particle swarm optimization (PSO) algorithm variant, population topology, and parameter settings, as these factors greatly affect the performance and efficiency of the optimization process. (NA?)

  • Utilise efficient algorithms for projecting a vector onto the (_{1})-ball, specifically the ones presented in the paper, to improve performance in high-dimensional learning tasks. (NA?)

  • Optimize a smooth proxy objective for non-smooth rank-based metrics, such as NDCG, in order to effectively train models for large complex ranking tasks. (NA?)

  • Consider utilizing the unique characteristics of bee swarms, including positive and negative feedback loops, fluctuations, multiple interactions, and division of labor, to inform the development of novel problem-solving techniques across various fields. (NA?)

  • Consider using stochastic methods for (_{1}) regularized loss minimization problems, such as the Lasso, as these methods offer significant advantages in terms of computational efficiency and scalability compared to deterministic approaches. (NA?)

  • Focus on developing online learning frameworks that utilize pairwise comparisons derived from implicit user feedback to improve information retrieval systems, rather than relying solely on traditional proxy-measures that may not accurately capture user utility. (NA?)

  • Consider implementing a system called Green’, which offers a simple and flexible framework allowing programmers to take advantage of approximation opportunities in a systematic manner while providing statistical QoS guarantees. (NA?)

  • Utilise the delta method’, a novel technique for capturing interacting variables in non-separable problems, which involves sorting decision variables based on the magnitude of your delta values and grouping those with smaller delta values together. (NA?)

  • Consider utilizing swarm intelligence methods like Ant Colony Optimization (ACO) and Particle Swarm Optimization (PSO) for effective search and data organization tasks in data mining projects. (NA?)

  • Utilise a majorization-minimization approach to tackle the sparse generalized eigenvalue problem, which provides a general, efficient algorithm to obtain sparse solutions to a wide range of scientific and engineering problems. (NA?)

  • Use coordinate descent methods to efficiently optimize the dual form of logistic regression and maximum entropy models, resulting in improved performance compared to traditional approaches. (NA?)

  • Consider using the Pegasos algorithm for solving the optimization problem cast by Support Vector Machines (SVM), as it offers a simple and effective iterative approach that alternates between stochastic gradient descent steps and projection steps, requiring fewer iterations compared to previous analyses of stochastic gradient descent methods. (NA?)

  • Utilize evolutionary algorithms (EAs) due to your assumption-free, flexible, robust, and exploratory nature, allowing for the discovery of innovative and effective solutions to complex problems. (NA?)

  • Consider utilising the Artificial Bee Colony (ABC) algorithm for efficient multivariate data clustering, as demonstrated through its successful application on thirteen typical test data sets from the UCI Machine Learning Repository, outperforming the Particle Swarm Optimization (PSO) algorithm and other nine classification techniques from the literature. (NA?)

  • Employ the Extended Local Similarity Analysis (eLSA) technique to effectively analyze time series data with replicates, enabling them to uncover statistically significant local and potentially time-delayed association patterns beyond those detectable via ordinary correlation analysis. (NA?)

  • Focus on developing robust learning algorithms, which are characterised by your ability to maintain similar levels of performance across testing and training samples that are close’, as this ensures better generalisation capabilities.’ (NA?)

  • Consider using incremental proximal methods for large scale convex optimization problems, as these methods offer advantages over traditional gradient and subgradient methods in terms of stability and potential ease of implementation. (NA?)

  • Utilise the Stochastic Alternating Direction Method of Multipliers (SADMM) algorithm for solving complex optimization problems involving stochastic and composite objective functions. (NA?)

  • Utilize CVXGEN, a software tool that automates the creation of custom C code for convex optimization problems, enabling efficient and reliable solutions for complex problem families. (NA?)

  • Carefully analyze the relationship between the convergence of your optimization algorithm and the underlying spectral properties of the network, as this can lead to improved understanding of network scaling issues and potentially better performance. (NA?)

  • Utilise a novel primal-dual splitting method for convex optimization involving Lipschitzian, proximable and linear composite terms, which provides a full splitting approach that processes individual components separately and avoids explicit inversions. (NA?)

  • Use hierarchical genetic algorithms which automatically allocate credit based on fitness levels, allowing for efficient exploration of solution spaces through the use of building blocks derived from high-fitness individuals. (NA?)

  • Utilize submodular functions in machine learning because they offer a powerful framework for expressing optimization problems and enable the development of efficient algorithms for a range of applications including clustering, experimental design, sensor placement, graphical model structure learning, and subset selection. (NA?)

  • Focus on developing hyper-heuristic algorithms that are more generally applicable than many current implementations of search methodologies, by finding the right method or sequence of heuristics in a given situation instead of trying to solve the problem directly. (NA?)

  • Use randomized block-coordinate descent (RBCD) methods for minimizing the sum of two convex functions, which can improve upon existing techniques like Nesterov (2012) and Richtarik and Takac (2014) by providing faster convergence rates and better high-probability type of iteration complexity. (NA?)

  • Utilize an inertial forward-backward splitting algorithm to efficiently compute a zero of the sum of two monotone operators, particularly when one of them is co-coercive. (NA?)

  • Utilize the randomized block-coordinate descent (RBCD) method for minimizing the sum of a smooth convex function and a block-separable convex function, as it provides a sharper expected-value type of convergence rate and a better high-probability type of iteration complexity compared to previous approaches. (NA?)

  • Consider using a conservative penalty term in your optimization problems to improve the convergence rate without sacrificing accuracy. (NA?)

  • Carefully select appropriate Gaussian processes priors and estimate your parameters accurately to achieve optimal convergence rates in efficient global optimization problems. (NA?)

  • Utilize the proposed algorithm for fast projection onto the simplex and the l1 ball, as it demonstrates superior speed compared to existing methods while still providing exact results in finite time. (NA?)

  • Consider utilizing parallel coordinate descent methods (PCDMs) when dealing with large-scale optimization problems involving partially separable composite objectives, as these methods offer significant speedups and improved efficiency compared to traditional serial methods. (NA?)

  • Consider utilizing the network lasso technique, which combines clustering and optimization in large graphs, to address complex optimization problems in various domains. (NA?)

  • Utilise the MISO’ (Minimisation by Incremental Surrogate Optimisation) scheme for efficient optimisation of large-scale machine learning problems, due to its low iteration costs and strong convergence guarantees even for nonconvex problems.’ (NA?)

  • Use SafeOpt, a Bayesian optimization algorithm, to optimize controller parameters for dynamic systems while maintaining safety and stability through modelling the performance measure as a Gaussian process and only exploring new controller parameters whose performance lies above a safe performance threshold with high probability. (NA?)

  • Carefully evaluate and choose the right machine learning algorithm and hyper-parameter values for your specific problem, considering factors like computational efficiency, ability to handle a wide range of algorithms, and capacity to manage variations in hyper-parameter numbers and types. (NA?)

  • Consider adopting a hybrid approach combining genetic algorithms (GA) and grey wolf optimization (GWO) for feature selection, followed by kernel extreme learning machines (KELM) for classification, particularly in medical diagnosis applications. (NA?)

  • Combine transfer learning techniques with scalable predictive modelling approaches to effectively handle large-scale, high-dimensional datasets in online advertising. (NA?)

  • Utilise diamond sampling’, a novel technique for estimating the maximum dot product between two sets of vectors, which is orders of magnitude faster than traditional methods and requires fewer samples. (NA?)

  • Utilise a novel methodology for generating datasets with varied appearances but identical statistical properties, achieved through simulated annealing optimisation strategies. (NA?)

  • Utilise a managed service for black-box optimization, which offers convenience and minimal user configuration, whilst hosting state-of-the-art algorithms and being highly scalable and adaptable. (NA?)

  • Carefully consider the choice of inertia weight, constriction factor, cognition and social acceleration coefficients, and topologies when applying the Particle Swarm Optimization algorithm to achieve optimal results. (NA?)

  • Focus on understanding the relationship between the population risk and the empirical distribution of parameters, as this enables them to analyze the PDE governing the distributional dynamics of stochastic gradient descent (SGD) in two-layer neural networks. (NA?)

  • Choose a suitable optimization framework based on your specific requirements, considering factors such as the programming language, ease of use, extendibility, and availability of relevant features such as parallelization, visualization, and decision-making tools. (NA?)

  • Carefully consider the choice of hyperparameter optimization technique, taking into account the characteristics of the machine learning model and the nature of the problem at hand, as different optimization techniques have varying strengths and limitations. (NA?)

  • Consider utilizing stochastic gradient methods such as SMD when dealing with large datasets and well-behaved functions to significantly reduce training times while maintaining good convergence properties. (NA?)

  • Consider implementing a multi-swarm particle swarm optimization (MSPSO) algorithm for solving complex optimization problems, such as feature selection, due to its ability to effectively balance exploration and exploitation while avoiding premature convergence. (NA?)

  • Utilise a coordinate descent algorithm for training linear SVM with the L2-loss function, as it provides a more efficient and stable alternative to current state-of-the-art methods like Pegasos and TRON. (NA?)

  • Consider using genetic programming (GP) in Inverse Generative Social Science (IGSS) for learning interpretable agent logic in agent-based models (ABMs) across various domains, as demonstrated by the successful application of this methodology in the fields of flocking and opinion dynamics. (NA?)

Stochastic Gradient Descent (Sgd)

  • Utilize the Krum aggregation rule in order to achieve Byzantine resilience in distributed Stochastic Gradient Descent (SGD) applications, thus enabling the system to continue functioning effectively even in the presence of malicious actors attempting to disrupt the process. (J. Feng, Xu, and Mannor 2017)

Artificial Intelligence Applications

  • Consider integrating large foundation models (such as LLMs and VLMs) into your agent-based AI systems to enhance your performance, adaptability, and responsiveness across diverse scenarios while addressing issues such as hallucinations and biases. (Durante et al. 2024)

  • Leverage large language models (LLMs) and prompt engineering techniques to achieve significant reductions in web accessibility violations, thereby contributing towards creating a more inclusive digital environment. (Calista Huang et al. 2024)

  • Leverage large language models (LLMs) to extract commonsense knowledge for planning to complete object rearrangement tasks, thereby enabling robots to better understand and respond to natural-language commands. (Y. Ding et al. 2023)

  • Leverage gamification techniques to increase user engagement and efficiency in collecting human-robot interaction (HRI) data, thereby improving the development and evaluation of generalizable and assistive embodied artificial intelligence (AI) systems. (Q. Gao et al. 2023)

  • Carefully evaluate the capabilities and limitations of ChatGPT for programming numerical algorithms across multiple programming languages, including its ability to handle debugging, code completion, code translation, and code parallelization, while being aware of potential issues like singular matrices, array compatibility, library inclusion, server disconnections, and default vs. plus versions. (Kashefi and Mukerji 2023)

  • Utilize multiple plagiarism detection tools and incorporate AI-based approaches like ChatGPT to effectively identify and prevent plagiarism in academic works. (Khalil and Er 2023)

  • Focus on investigating the potential of Artificial General Intelligence (AGI) in transforming education, considering its implications, challenges, ethics, and opportunities, especially in terms of personalized learning experiences, improved assessment methods, and the role of human educators in the face of advanced machine intelligence. (E. Latif et al. 2023)

  • Carefully investigate design choices to balance the alignment-fidelity tradeoff when fine-tuning text-to-image models using human feedback. (Kimin Lee et al. 2023)

  • Prioritize developing open-source, large language models for code (Code LLMs) that address copyright, privacy, transparency, and community-driven model development concerns, as demonstrated by the introduction of StarCoder and StarCoderBase models. (Raymond Li et al. 2023)

  • Consider introducing instruction tuning into multi-modal models, inspired by the Flamingo models upstream interleaved format pretraining dataset, to enhance your instruction-following ability and in-context learning.’ (Bo Li, Zhang, et al. 2023)

  • Focus on developing a comprehensive classification model for jailbreak prompts, which includes an iterative labeling process based on open coding methodology, to accurately categorize and analyze the effectiveness of jailbreak techniques in bypassing Large Language Model (LLM) restrictions. (Yi Liu et al. 2023)

  • Consider using MMBench, a systematically-designed objective evaluation benchmark, to robustly evaluate the various abilities of vision-language models, providing a comprehensive assessment of your performance across 20 different ability dimensions. (Yuan Liu et al. 2023)

  • Consider utilizing a two-stage prompt-based approach called PANDA (Prompt-based Context- and Domain-aware Pretraining) to effectively train vision-language models for VLN tasks. (T. Liu et al. 2023)

  • Focus on developing appropriate high-quality prompts to make large language models (such as ChatGPT and GPT-4) efficient and effective in protecting privacy, specifically in the context of medical text de-identification. (Zhengliang Liu et al. 2023)

  • Consider adopting a unified perspective of indirect supervision’ when analyzing different types of task instructions, as this allows for a deeper understanding of your advantages, limitations, and potential applications.’ (R. Lou, Zhang, and Yin 2023)

  • Focus on identifying and leveraging the unique properties of large language models (LLMs) for effective detection of machine-generated text, particularly utilizing the curvature of the models log probability function.’ (Mitchell et al. 2023)

  • Consider leveraging large language models (LLMs) to hypothesize an abstract world model (AWM) for reinforcement learning (RL) agents, which can then be verified through world experience, leading to improved sample efficiency and robustness against errors in the LLM. (Nottingham et al. 2023)

  • Aim to combine the strengths of specialized tools and foundation models to achieve enhanced accuracy, efficiency, and automation in problem-solving, while considering safety, trustworthiness, and personalization challenges. (Yujia Qin, Hu, et al. 2023)

  • Leverage large language models (LLMs) as controllers to manage and coordinate the efforts of various AI models in solving complex AI tasks, while using language as a generic interface to empower this connection. (Yongliang Shen et al. 2023)

  • Carefully consider your choice of prompting method when conducting Entity Resolution tasks using Large Language Models, as different approaches can impact both performance and cost. (Sisaengsuwanchai, Nananukul, and Kejriwal 2023)

  • Consider combining multimodal encoders from ImageBind and large language models from Vicuna to create a general-purpose model capable of instruction-following data from six modalities, allowing for improved cross-modal capabilities and a more comprehensive understanding of the world. (Yixuan Su et al. 2023)

  • Employ a unified framework for the architecture design of LLM-based autonomous agents, which includes a profiling module, a memory module, a planning module, and an action module, to ensure optimal performance across diverse tasks. (Lei Wang et al. 2023)

  • Use Point Prompt Training (PPT) to overcome negative transfer issues in multi-dataset synergistic learning for 3D representation learning, allowing for improved generalizability and performance across diverse datasets. (L. Wu et al. 2023)

  • Utilize a conversational APR approach for automated program repair, which combines patch generation and validation in a conversational manner, allowing the model to learn from previous incorrect patches and improve its accuracy over time. (C. S. Xia and Zhang 2023)

  • Leverage both Large Language Models (LLMs) and Answer Set Programming (ASP) technologies to create conversational agents that can effectively comprehend human dialogues and provide accurate responses based on a deeper understanding of the semantic meaning of sentences. (Y. Zeng et al. 2023)

  • Consider the importance of both general techniques (such as backbone architecture and self-supervised pretraining) and creation techniques (including likelihood-based models, energy-based models, and GANs) when developing generative AI systems for diverse content generation tasks. (Yi Zhang, Zhang, and Jiang 2023)

  • Consider adopting a prompt-based virtual assistant (VA) framework like BIM-GPT, which combines BIM and generative pre-trained transformer (GPT) technologies, to support natural language (NL)-based information retrieval (IR) from building information models (BIMs). (Junwen Zheng and Fischer 2023)

  • Carefully examine the performance of ChatGPT across various annotation tasks and labels, considering factors like precision, recall, and F1-score, to understand its strengths and limitations in generating human-like annotations. (Yiming Zhu et al. 2023)

  • Employ multiple methods, including quantitative-qualitative analysis, synthesis, abstraction, prediction, and experimental methods, to comprehensively evaluate the feasibility and effectiveness of implementing AI/ML-based prediction systems in public administration. (Ivashchenko, Ivashchenko, and Vasylets 2023)

  • Critically examine the accuracy, clarity, and potential bias of AI-generated information, specifically focusing on ChatGPTs ability to respond correctly and clearly to complex health-related topics such as COVID-19 vaccination conspiracies and compulsory vaccination.’ (Sallam et al. 2023)

  • Carefully consider the strengths and weaknesses of ChatGPT, a powerful AI tool, when incorporating it into your workflow, ensuring proper usage and verification of results to maximize its utility while minimizing potential biases and misinformation. (J.-J. Zhu et al. 2023)

  • Adopt a mixed-methods approach, combining both quantitative and qualitative data collection techniques, to thoroughly investigate the potential benefits and drawbacks of implementing ChatGPT in mathematics education. (Wardat et al. 2023)

  • Consider the limitations of ChatGPT when evaluating its performance in generating accurate and meaningful responses to complex chemistry assessment questions, especially those requiring visual interpretations or up-to-date information. (Fergus, Botha, and Ostovar 2023)

  • Consider integrating ChatGPT into your workflow for assistance in identifying new drug targets, generating new chemical structures, optimizing drug properties, assessing toxicity, and generating drug-related reports and papers, while acknowledging its limitations in performing complex computations and requiring experimental validation. (Rui Wang, Feng, and Wei 2023)

  • Consider the politics of automation, corporate infrastructure, scientization behind automation, conflicts over ethics, and the possibility of designing alternative solutions when studying the impact of AI, automation, and datafication in education. (Williamson, Macgilchrist, and Potter 2023)

  • Carefully examine the linguistic features of your subjects responses, specifically focusing on creativity, emotional expressiveness, and irreducible context-sensitivity, to determine if those subjects possess rational human mindedness, as opposed to relying solely on the apparent coherence or plausibility of the responses themselves.’ (Biever 2023)

  • Focus on developing Generalist Medical Artificial Intelligence (GMAI) models, which are capable of performing a diverse range of tasks using minimal labeled data, and can interpret different combinations of medical modalities, including imaging, electronic health records, laboratory results, genomics, graphs, or medical text. (Moor et al. 2023)

  • Carefully evaluate the potential opportunities and challenges presented by AI-driven code generation tools in order to effectively integrate them into your teaching practices and ensure optimal learning experiences for students. (Becker et al. 2023)

  • Use a minimalist preprocessing pipeline derived from MNE-Pythons default pipeline, followed by median evoked responses and standard single-trial linear decoding analyses, when working with MEG data due to its inherent noise issues.’ (Gwilliams et al. 2023)

  • Consider integrating ChatGPT, a natural language processing (NLP) model powered by OpenAIs GPT-3 technology, into your studies to enhance e-commerce via chat, as well as other sectors such as education, entertainment, finance, health, news, and productivity, by analyzing current use-cases and exploring possible future applications.’ (“A Review of ChatGPT AI’s Impact on Several Business Sectors” 2023)

  • Consider utilizing ChatGPT as an adjunct informational tool for patients and physicians to improve outcomes in the management of cirrhosis and HCC, despite its limitations in specifying decision-making cut-offs and treatment durations. (Yeo et al. 2023)

  • Utilise ChatGPT, an artificial intelligence tool, to generate suggestions for improving clinical decision support (CDS) logic, as it offers unique perspectives, is highly understandable and relevant, and can help identify potential improvements to alert logic and support your implementation. (Siru Liu et al. 2023)

  • Carefully examine the performance of large language models (LLMs) like ChatGPT and GPT-4 on specialized medical board examinations, taking into account factors such as question complexity, subspecialty area, and the presence of higher-order problem-solving skills. (R. Ali et al. 2023)

  • Utilize generative network models actively accounting for non-independence of data points via correlated random effects at both the node and dyad level during the process of model fitting, rather than relying solely on post-hoc permutation methods. (Ross, McElreath, and Redhead 2022)

  • Leverage Generative Pre-trained Transformer (GPT) models, specifically ChatGPT, for analyzing and generating text, as they significantly outperform traditional Natural Language Processing (NLP) techniques in accurately interpreting and providing reasoning for complex language, such as Fedspeak. (Akyürek et al. 2022)

  • Utilize DendroMap, a novel interactive visualization system for exploring large-scale image datasets used in machine learning, which adapts an interactive zoomable treemap and supports the information seeking mantra “overview first, zoom and filter, then details-on-demand”. (Bertucci et al. 2022)

  • Carefully consider whether weighted prompt engineering is helpful or hindering in your specific context, taking into account its potential impact on data quality and interpretation. (R. Gal et al. 2022)

  • Adopt a mixed-methods approach combining literature reviews, tool evaluations, and expert interviews to understand the principles, components, roles, and architecture of Machine Learning Operations (MLOps) and ultimately improve the automation and operation of ML products. (Kreuzberger, Kühl, and Hirschl 2022)

  • Consider reframing forecasting as future object detection, enabling the development of end-to-end models capable of predicting multiple future trajectories directly from LiDAR data, ultimately improving overall accuracy and encouraging a reevaluation of the role of explicit tracking in embodied perception. (Peri et al. 2022)

  • Investigate the potential benefits and challenges associated with integrating generative AI, such as ChatGPT, into engineering education, considering factors such as bias, misinformation, ethics, and employment implications, in order to optimize its positive impact while minimizing negative consequences. (Susnjak 2022)

  • Strive to develop machine learning sensors that adhere to the “Sensor 2.0” paradigm, which emphasizes modularity, data centricity, simplicity, transparency, and openness in order to overcome the limitations of the current approach to integrating machine learning into embedded systems. (Warden et al. 2022)

  • Employ a combination of quantitative and qualitative approaches when investigating the impact of ChatGPT on the lecturer profession, considering both the benefits and challenges posed by the technology while remaining vigilant against potential misuse. (Ausat 2022)

  • Learn from the experience of the chess world in adapting to AI advancements, finding the right balance between embracing the benefits of AI like ChatGPT while maintaining transparency, credibility, and combatting potential issues like plagiarism and misuse. (Stokel-Walker 2022)

  • Carefully consider the implications of using learned sparse representations in information retrieval tasks, as these models can lead to counter-intuitive term weight assignments that negatively impact query performance and necessitate the need for tailored optimization techniques. (Mackenzie, Trotman, and Lin 2022)

  • Carefully define and categorize your contributions within the broader framework of AIOps, specifically focusing on failure management, to ensure clear communication and comparability among various approaches. (Notaro, Cardoso, and Gerndt 2021)

  • Use active learning techniques to efficiently generate seed alignment data for neural entity alignment tasks, taking into account both structural dependencies among entities and the presence of “bachelor” entities without counterparts in the other knowledge graph. (Bing Liu et al. 2021)

  • Utilize the CodeXGLUE benchmark dataset to facilitate machine learning research for program understanding and generation, providing a comprehensive platform for model evaluation and comparison, along with three baseline systems for ease of use. (S. Lu et al. 2021)

  • Consider developing a general-purpose vertical end-to-end machine learning platform, such as Looper, to efficiently implement data-driven real-time smart strategies in various domains, while incorporating causal product-impact evaluations and optimization, handling heterogeneous treatment effects, and utilizing meta-learning algorithms. (Markov et al. 2021)

  • Utilise the newly introduced SciOL and MuLMS-Img datasets to improve the performance of large-scale vision-language models in the scientific domain, particularly for image-text tasks like figure type classification, optical character recognition, captioning, and figure retrieval. (Z. Shen et al. 2021)

  • Carefully consider the assumptions underlying your chosen metrics, particularly in relation to the handling of missing ratings and the conditional nature of item aspects on users, in order to accurately capture both diversity and accuracy in top-n recommendations. (Parapar and Radlinski 2021)

  • Focus on developing low-cost, non-invasive, compatible, ubiquitous, and flexible occupancy detection systems using Bluetooth Low Energy (BLE) technology, rather than relying on expensive and less versatile alternatives such as WiFi-based systems. (Demrozi et al. 2021)

  • Use the population average prescriptive effect (PAPE) and the area under the prescriptive effect curve (AUPEC) as evaluation metrics for individualized treatment rules (ITRs), as they provide a comprehensive assessment of the performance of ITRs while taking into consideration the proportion of units treated and budget constraints. (Imai and Li 2021)

  • Utilize supervised machine learning techniques, specifically “learning to rank” algorithms, to accurately locate delivery points for addresses using noisy GPS data. (“Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track” 2021)

  • Focus on modeling user interaction in social media sessions to enhance the accuracy of cyberbullying detection, specifically by analyzing the temporal dynamics and topic coherence of user comments. (S. Ge, Cheng, and Liu 2021)

  • Employ a combination of burst detection and statistical probing methods to accurately identify smart home devices that unexpectedly record and send audio to the internet, thereby improving overall detection accuracy. (Mitev et al. 2020)

  • Utilise a combination of relevance and diversity measures in your search algorithms to ensure optimal results for users. (Abdool et al. 2020)

  • Focus on developing a robust and efficient control system for managing online advertising budgets, taking into account factors like plant latency, excitation latency, and integral gain, while ensuring stability and minimizing overshoots and oscillations. (Karlsson 2020)

  • Consider multiple factors beyond just relevance when evaluating search results, including personalized preferences, popularity, exploration mode, tolerance, price, and indirect relation, as these factors can impact user engagement and purchase decisions. (Carmel et al. 2020)

  • Focus on building customs fraud detection models that are highly accurate, interpretable, and able to handle large amounts of data while considering factors such as data quality, privacy concerns, and changing patterns of fraud. (Sundong Kim et al. 2020)

  • Explore multi-source (multimodal) methods for AIOps by combining several data source categories like metrics, logs, and traces to overcome the limitations of single observer and single method approaches, leading to improved accuracy and reduced false positives in fault detection, root-cause analysis, and remediation. (Nedelkoski et al. 2019)

  • Use prompt engineering techniques such as zero-shot learning, few-shot learning, chain-of-thought prompting, and ask-me-anything prompting to improve the mathematical reasoning capabilities of large language models like GPT-4, thereby enhancing your ability to provide accurate solutions for complex mathematical problems. (Lample and Charton 2019)

  • Engage in threat modeling to identify potential threats and vulnerabilities associated with neural fake news, and subsequently develop robust defenses against them. (Zellers, Holtzman, Rashkin, et al. 2019)

  • Focus on observing and analyzing the complex interactions among various actors and systems in real-world scenarios, particularly in terms of anomaly response, in order to identify patterns and vulnerabilities that can inform the development of effective strategies and tools for improving performance. (S. Chuang et al. 2019)

  • Utilise a combination of active learning and advanced search techniques, such as Approximate Nearest Neighbour (ANN) search and Maximum Inner Product Search (MIPS), to effectively and efficiently predict click-through rates (CTR) for billions of user query and ad pairs in sponsored search systems. (M. Fan et al. 2019)

  • Consider implementing Long Short-Term Memory (LSTM) networks in memristor crossbars to overcome limitations in computing power due to limited memory capacity and data communication bandwidth, thereby enhancing the potential of these networks for use in edge inference. (Smagulova, Krestinskaya, and James 2018)

  • Develop a novel Graph-based Causal Inference (GCI) framework to enable effective causal inference in unstructured data, particularly in legal domains, through building causal graphs from fact descriptions without significant human input, enabling accurate decision-making. (Devlin et al. 2018)

  • Carefully consider the choice of regularization and the formulation of the objective function when working with tensor factorization methods for knowledge base completion. (Lacroix, Usunier, and Obozinski 2018)

  • Employ stochastic geometry models to analyze the average performance of NOMA-aided UAV networks, taking into account factors like user pairing strategies, multiple-UAV cases, and power allocation schemes. (Yuanwei Liu et al. 2018)

  • Utilize the FEVER dataset for developing and testing claim verification algorithms against textual sources, as it provides a comprehensive and diverse collection of claims and evidence, enabling accurate assessments of algorithm performance. (Thorne et al. 2018)

  • Utilise a unified framework for personal comfort models, which involves determining suitable data for learning algorithms, processing and preparing raw data, selecting appropriate learning algorithms, validating predictive performance, and updating the model based on new data to ensure accuracy and relevance over time. (Joyce Kim, Schiavon, and Brager 2018)

  • Develop a comprehensive fire reporting system that utilizes advanced technologies such as smoke, carbon monoxide, and temperature sensors, along with a reliable communication module, to enable rapid and accurate detection and notification of fires, thereby reducing potential damage and loss. (Aiyelabowo et al. 2018)

  • Consider combining IoT-based sensors, big data processing, and machine learning models like DBSCAN and Random Forest to create comprehensive monitoring systems for industries like automotive manufacturing. (Syafrudin et al. 2018)

  • Leverage the vast amount of open-source software data available to develop machine learning algorithms that can identify patterns in well-written code, thereby improving software reliability, readability, and maintainability. (Allamanis et al. 2017)

  • Carefully consider the potential sources of bias and measurement errors when analyzing the relationship between artificial intelligence and productivity growth, including factors such as false hopes, mismeasurement, redistribution, and implementation lags. (Fortunato et al. 2017)

  • Utilise a deep evolutionary knowledge network called “Know-Evolve”, which employs a multivariate point process to model the occurrence of facts in dynamic knowledge graphs, thus improving the accuracy of predictions related to the occurrence or recurrence time of a fact. (Trivedi et al. 2017)

  • Prioritize prediction accuracy over statistical significance when evaluating the validity of your models, as prediction allows for a more comprehensive assessment of the models overall fit and ability to inform policy decisions.’ (Hegre et al. 2017)

  • Carefully select appropriate sEMG acquisition setups for your specific requirements, taking into account factors like cost, robustness, and ease of use, while ensuring comparability across studies. (Pizzolato et al. 2017)

  • Utilize a minimax game-based approach to unify generative and discriminative information retrieval (IR) models, allowing them to continuously challenge and improve each other, leading to significant improvements in performance across multiple IR applications. (Jun Wang et al. 2017)

  • Consider using high-level synthesis (HLS) techniques like LegUp to develop efficient and scalable FPGA-based CNN accelerators, leveraging the benefits of software parallelism and precise tailoring to the target CNN. (J. H. Kim et al. 2017)

  • Utilize a combination of ranking functions, semantic matching features, and query rewriting to optimize search engine relevance, particularly for tail queries. (D. Yin et al. 2016)

  • Combine machine-learning algorithms with crowdsourcing techniques to efficiently and accurately assess media bias at scale, allowing them to overcome the limitations of existing content-based methods. (Budak, Goel, and Rao 2016)

  • Consider using a Multi-Perspective Context Matching (MPCM) model when working with the SQuAD dataset for machine comprehension tasks, as it effectively combines multiple matching strategies to improve overall performance. (Zhiguo Wang et al. 2016)

  • Consider using a combination of dense range images and sparse convolutions to create a more efficient and accurate 3D object detection model for autonomous driving systems. (J. L. Ba, Kiros, and Hinton 2016)

  • Focus on developing algorithms that leverage cryptographic sortition and secret credentials to create a decentralized, tamperproof ledger system that ensures fair representation and prevents manipulation by malicious actors. (Jing Chen and Micali 2016)

  • Utilise Pixel Recurrent Neural Networks (PixelRNNs) when modelling the distribution of natural images due to your ability to accurately predict pixels sequentially, thereby improving the overall accuracy of the model. (Oord, Kalchbrenner, and Kavukcuoglu 2016)

  • Consider combining telecommunications data with land cover data to improve the accuracy of population estimates, particularly through the use of random forest regression techniques. (Douglass et al. 2015)

  • Carefully consider the choice of distance metric when comparing point sets, as different metrics capture different aspects of shape space and can impact the overall accuracy of the generated 3D point clouds. (A. X. Chang et al. 2015)

  • Consider utilising multiresolution tree-structured networks when working with 3D point cloud processing tasks, as they offer advantages such as efficient feed-forward processing, coarse-to-fine analysis, faster convergence, and smaller memory footprint during training. (A. X. Chang et al. 2015)

  • Utilize a comprehensive, user-centered framework for assessing and comparing AutoML services, incorporating six distinct categories - Estimates, Scope, Productivity, Interpretability, Customizability, and Connectivity - to ensure a holistic evaluation of the services effectiveness and suitability for diverse analytic tasks.’ (Laptev, Amizadeh, and Flint 2015)

  • Consider combining learning, rules, crowdsourcing, in-house analysts, and developers as “first-class citizens” in your study designs, especially when tackling large-scale classification problems. (C. Sun et al. 2014)

  • Consider the challenges posed by urban sensing and data acquisition, computing with heterogeneous data, and hybrid systems blending the physical and virtual worlds when conducting urban computing projects. (Y. Zheng et al. 2014)

  • Focus on developing automated solutions for optimizing bidding strategies in real-time bidding (RTB) auctions, considering various factors such as campaign-specific key performance indicators (KPIs), campaign lifetime auction volumes, and budgets. (Weinan Zhang et al. 2014)

  • Consider using deep learning techniques when analyzing complex datasets, particularly those involving multiple feature extraction methods like static and dynamic analyses, as they can lead to higher levels of accuracy compared to traditional machine learning models. (Zhenlong Yuan et al. 2014)

  • Consider implementing a budget pacing algorithm in your ad serving systems to improve overall performance and satisfaction for both advertisers and publishers. (D. Agarwal et al. 2014)

  • Consider incorporating a secondary model to capture the expected delay between a click and a conversion in order to improve the accuracy of conversion predictions in display advertising. (Chapelle 2014)

  • Utilize a novel functional optimization framework to determine the optimal bidding strategy in RTB display advertising, taking into consideration the budget constraint, campaigns lifetime, probability of winning the auction, and the prior distribution of the impression features.’ (Weinan Zhang, Yuan, and Wang 2014)

  • Use multi-touch attribution instead of last-touch attribution for accurate budget allocation in online advertising, as it enables a more precise understanding of the influence of various sub-campaigns on consumer actions. (Geyik, Saxena, and Dasdan 2014)

  • Utilise the iPinYou Global RTB Bidding Algorithm Competition Dataset to conduct experiments on important issues in computational advertising such as bid optimisation and CTR estimation, due to its status as the first publicly available dataset on RTB display advertising. (H. Liao et al. 2014)

  • Use a combination of manual inspection, automated tools, and multiple data sources to accurately identify and track Bitcoin-related scams, including obtaining information from online forums, blockchain analysis, and historical snapshots of websites. (Vasek 2014)

  • Utilise a multimodal contrastive approach to connect languages via visual observations, allowing for accurate machine translation without the need for parallel corpora. (El-Shishtawy and El-Sammak 2014)

  • Be aware of potential information leakage from machine learning classifiers, as it is possible to infer unexpected but useful information from them, which can lead to security risks and intellectual property violations. (Ateniese et al. 2013)

  • Optimize interleaving algorithms for online retrieval evaluation by solving an optimization problem subject to certain constraints, resulting in an unbiased and more efficient algorithm than previously developed approaches. (Radlinski and Craswell 2013)

  • Focus on developing a formal model for peer-to-peer communication and the Proofs of Work concept used in Bitcoin, and demonstrate how standard primitives from distributed computation, like broadcast and MPCs, can be implemented in this model. (Bahack 2013)

  • Consider utilizing a parameter server framework for distributed machine learning problems, as it effectively addresses issues like network bandwidth consumption, synchronization barriers, and fault tolerance through its asynchronous communication model, flexible consistency models, elastic scalability, and continuous fault tolerance. (Xing et al. 2013)

  • Carefully select appropriate evaluation metrics for predictive models in social media personality prediction studies, as certain metrics like Mean Average Error (MAE) and Root Mean Square Error (RMSE) might hide inaccuracies when attempting to predict extreme percentiles. (Sumner et al. 2012)

  • Consider utilizing the open-source Python library Scikit-learn’, which offers a comprehensive suite of machine learning algorithms, emphasizes ease of use, performance, and API consistency, and is designed specifically for non-specialist users.’ (Pedregosa et al. 2012)

  • Utilise the Million Song Dataset (MSD) to evaluate and improve your algorithms, particularly for tasks requiring large-scale data, due to its size, diversity, and linkage to multiple resources. (Bertin-Mahieux et al. 2011)

  • Focus on developing models that accurately predict the difference in response rates between treatment and control groups rather than simply predicting the overall response rate in either group. (Rzepakowski and Jaroszewicz 2011)

  • Consider using parallel computing methods like the proposed parallel boosted regression trees (PBRT) algorithm to efficiently handle large datasets in web search ranking applications, resulting in almost perfect linear speed-ups and minimal loss in accuracy. (Tyree et al. 2011)

  • Conduct a systematic mapping study (SMS) to gain a comprehensive understanding of the AIOps field, involving careful formulation of research questions, thorough search and selection processes, and rigorous data extraction and categorization. (Abreu, Zoeteweij, and Gemund 2009)

  • Aim to create a universally applicable, formalized definition of machine intelligence that does not rely on specific sets of senses, environments, or hardware, and that is based on fundamental principles unlikely to change over time. (Legg and Hutter 2007)

  • Adopt a comprehensive, multi-step approach to knowledge discovery in databases (KDD), which goes beyond mere data mining and includes crucial stages such as data preparation, data selection, data cleaning, incorporation of prior knowledge, and proper interpretation of results. (Xindong Wu et al. 2007)

  • Carefully consider the impact of communication constraints on the design of learning algorithms for wireless sensor networks, as traditional centralized learning strategies may not be feasible due to energy and bandwidth limitations. (Predd, Kulkarni, and Poor 2006)

  • Develop a web-based application called iSTART, which provides young adolescent to college-age students with high-level reading strategy training to improve comprehension of science texts, based on the principles of self-explanation reading training (SERT) and incorporating various interactive elements to facilitate learning. (McNamara, Levinstein, and Boonthum 2004)

  • Utilise dynamic item response models and Bayesian methods to accurately capture the potential variability of ideal points over time, rather than assuming them to be static. (Martin and Quinn 2002)

  • Prioritize simplicity and interpretability in your choice of logical rules, opting for crisp rules initially and only resorting to fuzzy rules or more advanced methods if necessary, while considering the trade-offs between accuracy, rejection levels, and complexity throughout the process. (Duch, Adamczak, and Grabczewski 2001)

  • Consider utilizing a Contract Definition Language (CDL) to enable automated reasoning over legal substance, thereby increasing the efficiency and effectiveness of contract formulation, analysis, and execution. (L. E. Allen and Saxon 1995)

  • Focus on developing practical implementations of public-key cryptosystems, specifically through the creation of trap-door one-way permutations, to ensure privacy and enable digital signatures in electronic communication. (Rivest, Shamir, and Adleman 1978)

  • Consider adopting a hierarchical Bayesian approach for location estimation in wireless networks, as it enables accurate predictions without the need for extensive training data or detailed maps of the environment. (Madigan et al., n.d.)

  • Collaborate closely with industry to create online platforms and physical spaces for solving real-world industrial problems, while developing protocols for data privacy, protection, and security. (NA?)

  • Consider moving away from traditional objectivist views of learning towards a more constructivist approach, recognising that individuals construct your own unique versions of reality based on your experiences and interpretations. (NA?)

  • Utilize case-based reasoning (CBR) to effectively address complex problem-solving scenarios by drawing upon relevant past experiences, adapting existing solutions to fit new challenges, and continuously refining your approach through evaluation and repair. (NA?)

  • Consider the four dimensions of classification for adaptive hypermedia methods and techniques: application areas, user features, technologies of adaptation, and adaptation goals. (NA?)

  • Utilise a hybrid user modelling strategy when developing intelligent information agents, incorporating both short-term and long-term user interests, and tracking information already presented to the user. (NA?)

  • Utilize a multi-strategy machine learning approach when developing intelligent information agents, allowing for the induction of user models that consist of separate models for long-term and short-term interests, and employ “concept feedback,” a novel form of user feedback that enables users to critique the agents explanations for its classifications, thereby permitting more direct modifications to the induced concepts than through the inclusion of additional training examples.’ (NA?)

  • Consider multiple factors beyond just accuracy when comparing machine learning techniques for software effort prediction, including explanatory value and configurability. (NA?)

  • Adopt standardized nomenclature and compatible modelling techniques to facilitate the integration of independent functional models into a comprehensive model of the cell under study. (NA?)

  • Focus on developing direct brain-machine interfaces (BMIs) that utilize intracortical recording to improve the speed, flexibility, and accuracy of neural signal translation into command signals for controlling devices, ultimately aiming towards creating a comprehensive BMI solution for paralyzed individuals. (NA?)

  • Utilize both supervised and unsupervised statistical methods for fraud detection, considering factors such as uneven class sizes, different costs of misclassifications, and the costs of investigating observations, while also recognizing the limitations of statistical analysis in definitively proving fraud. (NA?)

  • Utilize Support Vector Machines (SVMs) with a binary tree recognition strategy for audio classification tasks, and employ a new metric called Distance-From-Boundary’ (DFB) for audio retrieval, as these methods offer significant improvements over existing techniques.’ (NA?)

  • Consider multiple dimensions of performance metrics, including recall and precision, using target and lexical criteria for page relevance, and assess the efficiency of the crawling algorithms, to provide a comprehensive understanding of the performance of web crawlers. (NA?)

  • Strive to create a unified theory that bridges the gaps between cognition and affect, and between theory and practice, in order to achieve a balanced and comprehensive understanding of the science of learning. (NA?)

  • Utilise a combination of machine learning techniques, specifically a naive Bayes model and a discriminative model, to effectively disambiguate author names within citation databases. These models allow for the consideration of various factors such as co-authorship patterns, paper titles, and journal titles, thereby increasing the accuracy of name identification and reducing errors caused by name ambiguities. (NA?)

  • Consider the potential impact of cortical representations of artificial actuators on neural adaptation and optimization of neuronal representation of new behavioral goals. (NA?)

  • Prioritise the development of algorithms that intelligently select the number of Access Points (APs) used for location estimation, thereby enabling increased estimation accuracy whilst simultaneously reducing power consumption. (NA?)

  • Participate in BCI competitions to validate various data analysis techniques and improve the reliability of brain-computer interfaces. (NA?)

  • Consider the various components involved in creating a successful medical decision support system (DSS), including the targeted decision-makers, user interface, data, algorithms and tools, and additional resources, while keeping in mind the ethical implications and challenges posed by the widespread use of DSS. (NA?)

  • Utilise multiple social psychology behavioural and decision-making models when studying the adoption of domestic robots, taking into account both personal and social factors, as well as the unique challenges posed by these technologies. (NA?)

  • Consider using Bayesian methods to analyze how individuals optimally combine multiple sources of information, including current and prior experiences, to make accurate judgments and navigate complex environments. (NA?)

  • Aim to create a universally applicable, formalized definition of machine intelligence that incorporates the core elements of human intelligence, is independent of specific senses, environments, or hardware, and is capable of effectively evaluating a wide variety of systems. (NA?)

  • Consider integrating the functionalities of Weka and R through the RWeka package, which enables seamless interaction between the two platforms, thereby facilitating efficient data pre-processing, exploratory analysis, and model fitting in a single statistical environment. (NA?)

  • Consider using machine learning techniques, specifically Kernel Canonical Correlation Analysis (KCCA), to predict multiple performance metrics of database queries using only information available prior to query execution. (NA?)

  • Consider utilizing the Computing with Words (CW) methodology when dealing with decision making problems involving imprecise, uncertain, or partial information, as it enables a more human-like approach to modeling perceptions and preferences. (NA?)

  • Develop a probabilistic model of user affect that combines both causal and diagnostic information to accurately assess users emotional states during interactions with educational computer games.’ (NA?)

  • Consider combining simulation and optimization modeling approaches when studying reservoir systems operations, as it can lead to better results than using either technique alone. (NA?)

  • Use the LETOR benchmark collection to develop and compare learning to rank algorithms for information retrieval, as it provides standardized document corpora, query sets, features, and evaluations, making it easier to develop algorithms and compare results across studies. (NA?)

  • Consider breaking down the complex process of evaluating interactive adaptive systems into smaller, more manageable parts, allowing for a deeper understanding of the underlying mechanisms and facilitating improvements in the design and functionality of these systems. (NA?)

  • Consider integrating advanced technologies like real-time computer graphics, virtual and augmented reality, and artificial intelligence into the development of serious games for cultural heritage purposes, as these technologies can enhance the accessibility, engagement, and educational effectiveness of such games. (NA?)

  • Focus on understanding and utilizing the computational power of physical bodies, specifically through the concept of morphological computation, to effectively simplify complex tasks and improve overall performance in various fields. (NA?)

  • Utilize the Nipype framework, an open-source, community-developed, Python-based software package, to efficiently analyze neuroimaging data and develop algorithms for comparative purposes. This framework addresses various challenges faced by researchers in the field, including lack of interoperability between different neuroimaging software packages, difficulty in reproducibility of results, and limited scalability for handling large datasets. (NA?)

  • Consider using the Online Chemical Modeling Environment (OCHEM) platform for performing QSAR/QSPR studies, as it simplifies the modeling process, enables data sharing, and promotes collaboration within the scientific community. (NA?)

  • Carefully consider the type of transfer learning they are attempting, whether it involves crossing sensor boundaries, dealing with physical setting differences, or working with limited labeled data, and choose appropriate methods accordingly. (NA?)

  • Utilize metamorphic testing to validate and verify machine learning classifiers, especially in situations where traditional test oracles are absent or unavailable. (NA?)

  • Leverage large-scale smartphone data to explore the relationship between automatically extracted behavioral characteristics and self-reported Big-Five personality traits, while considering potential gender-based differences in the analysis. (NA?)

  • Utilise big mobility data to bridge the gap between raw data and meaningful insights, by leveraging the sheer size and precision of the data to uncover patterns and trends in human mobility behaviour. (NA?)

  • Consider employing quantitative EEG (qEEG) alongside traditional EEG methods to improve the accuracy and efficiency of detecting ischemic changes in various clinical situations, especially in cases where raw EEG might miss subtle changes or be too time-consuming to analyze. (NA?)

  • Consider utilizing the open-source Python library Scikit-learn’, which offers a comprehensive suite of machine learning algorithms, emphasizes usability, performance, and documentation, and is designed specifically for non-specialist users.’ (NA?)

  • Carefully consider the type of feedback being utilised (whether its visual, auditory, haptic, or multimodal), the complexity of the motor task, and the stage of learning when designing feedback interventions for motor learning.’ (NA?)

  • Utilize the Push programming language in your evolutionary computation systems, as it enables the automatic provision of advanced genetic programming facilities (multiple data types, automatically defined subroutines, control structures, and architecture) without requiring extra machinery or user configuration. (NA?)

  • Utilize a deep multi-task artificial neural network when developing a machine learning model for predicting multiple electronic ground- and excited-state properties in organic molecules, taking advantage of the underlying correlations between various molecular properties. (NA?)

  • Focus on utilizing the crystalline state of GST, which differs significantly from previous approaches, to achieve high-speed synaptic events and efficient memory storage. (NA?)

  • Seek access to open-source “big” (social media) data sets and non-programming interfaces for deep data analysis, while addressing the challenges of data cleansing, holistic data sources, data protection, sophisticated data analytics, analytics dashboards, and data visualization in your studies. (NA?)

  • Prioritize the empirical study of users in interactive machine learning systems, focusing on your interactions, behaviors, and needs, to enhance the efficiency and effectiveness of these systems. (NA?)

  • Consider utilizing the scikit-learn Python library for neuroimaging data analysis, as it provides a comprehensive suite of machine learning algorithms that can effectively handle high-dimensional datasets, enabling accurate modeling of brain activities and behaviors. (NA?)

  • Carefully evaluate the advantages and disadvantages of each machine learning algorithm before selecting the appropriate solution for your specific wireless sensor network challenge. (NA?)

  • Consider the importance of addressing transitions and unknown activities in Human Activity Recognition (HAR) systems, as ignoring them may lead to decreased system accuracy and functionality. (NA?)

  • Consider implementing a quantum support vector machine (QSVM) for big data classification tasks, as it offers a significant improvement in computational efficiency compared to traditional methods, particularly in cases where classical sampling algorithms require polynomial time. (NA?)

  • Combine human and machine intelligence to efficiently classify crisis-related microblog communications into user-defined categories, allowing for improved disaster response efforts. (NA?)

  • Recognize the importance of prediction problems alongside causal inference problems, and leverage machine learning techniques to optimize prediction accuracy in policy contexts. (NA?)

  • Use machine learning algorithms to develop predictive models of brain age based on neuroimaging data, and then compare the predicted brain age of TBI patients to your actual age to identify potential acceleration of brain atrophy. (NA?)

  • Prioritize the creation of a reliable and representative dataset for accurate comparisons and validation of DPI-based traffic classification tools. (NA?)

  • Consider utilizing a deep learning architecture with a zero-masking strategy for data fusion to improve the accuracy of Alzheimers disease diagnosis by integrating multi-modal neuroimaging features.’ (NA?)

  • Use the Weka machine learning workbench to explore various machine learning algorithms and data pre-processing methods for solving common data mining problems in bioinformatics research, such as classification, regression, clustering, and feature selection, while comparing different techniques on the same problem to identify the most suitable algorithm for generating an accurate predictive model. (NA?)

  • Consider employing multiple regression methods when analyzing soil properties using VIS-NIR spectroscopy, as different methods may perform better depending on the specific soil property being investigated. (NA?)

  • Consider using a blended model that combines multiple feature sets, such as Bag of Words, hateful terms, and typed dependencies, to improve the classification performance of cyber hate speech detection systems, particularly in cases where intersectionality is involved. (NA?)

  • Consider utilizing convolutional neural networks for the classification of hand movements in prosthetic devices, as they demonstrate superior performance compared to classical machine learning techniques. (NA?)

  • Utilize machine learning algorithms for attack detection in the smart grid, specifically focusing on supervised, semi-supervised, decision and feature level fusion, and online learning algorithms, as they outperform state vector estimation methods in detecting both observable and unobservable attacks. (NA?)

  • Employ an adaptive design approach in materials discovery, which involves utilizing uncertainties and maximizing the “expected improvement” from the best-so-far material in an iterative loop with feedback from experiments, effectively balancing the goal of searching for materials likely to have the best property with the need to explore parts of the search space with fewer sampling points and greater uncertainty. (NA?)

  • Consider developing a general-purpose attribute set for materials science, which can be adapted to a wide range of materials problems, thus simplifying the process of creating accurate machine learning models. (NA?)

  • Consider applying machine learning techniques to address the growing complexity and dynamism in manufacturing systems, particularly focusing on selecting appropriate algorithms and interpreting results while considering the unique characteristics of the available data and the specific goals of the study. (NA?)

  • Utilise the Rayyan application to streamline and enhance the efficiency of the systematic review process, especially in the initial stages of screening abstracts and titles, thereby saving valuable time and improving the overall quality of the review. (NA?)

  • Consider utilizing data-driven techniques in disaster information management to enhance situation awareness, address users information needs, and apply data mining and machine learning techniques to improve overall disaster management efficiency.’ (NA?)

  • Combine crowdsourced human classification with machine learning algorithms to efficiently and accurately categorize complex data like glitches in LIGO detectors, ultimately improving the rate and accuracy of gravitational-wave observations. (NA?)

  • Utilize deep learning techniques, specifically deep convolutional neural networks, for improved robustness and efficiency in image-based plant phenotyping tasks. (NA?)

  • Consider using a semantic low-code engineering framework, such as SeLoC-ML, to enable efficient and scalable development of machine learning applications in industrial IoT settings. (NA?)

  • Focus on studying cross-device search behavior, specifically characterizing transitions between devices, to develop accurate prediction models for improving search support across multiple devices. (NA?)

  • Develop an automated testing framework to generate reproducible experiments with traffic generated by real applications, enabling evaluation of how these apps interact with Multipath TCP. (NA?)

  • Consider implementing a Proportional Integral Derivative (PID) based feedback-control system to manage multiple Key Performance Indicators (KPIs) in real-time bidding (RTB) display advertising, which involves generating a control score for each KPI based on the output of a PID controller module and a metric that quantifies the importance of each KPI for internal business needs, and choosing the KPI with the greatest overall need for improvement on regular intervals (NA?)

  • Carefully choose the right level of granularity for your fingerprints or descriptors, balancing prediction accuracy with computational efficiency, and use rigorous statistical practices such as cross-validation and testing on unseen data to avoid overfitting. (NA?)

  • Consider utilizing machine learning techniques, specifically Random Forest Regression (RFR), alongside multiple linear regression (MLR) for enhanced accuracy in predicting soil properties using remote sensing data. (NA?)

  • Consider using the MyoGym dataset, which contains 6D motion signals and 8 channel electromyogram data from 10 individuals performing 30 different gym exercises, for developing activity recognition classifiers, creating models for unseen activities, and conducting signal fusion studies. (NA?)

  • Employ an attention-based neural network method to jointly model the charge prediction task and the relevant article extraction task in a unified framework, allowing for improved accuracy in determining appropriate charges for a given case. (NA?)

  • Consider utilizing quantum feature spaces for machine learning tasks, particularly when dealing with large feature spaces or computationally expensive kernel functions, as this could potentially offer significant computational advantages. (NA?)

  • Focus on developing a new global Gross Primary Productivity (GPP) dataset called VPM GPP V20, which is based on an improved Light Use Efficiency (LUE) theory that considers the energy absorbed by chlorophyll, and applies this globally using remotely sensed datasets along with reanalysis climate datasets and land cover classifications. (NA?)

  • Utilise convolutional neural networks (CNNs) to accurately identify and locate quantum phase transitions in quantum many-fermion systems, even those exhibiting a severe fermion sign problem, by leveraging auxiliary-field quantum Monte Carlo (QMC) simulations to sample the many-fermion system. (NA?)

  • Carefully consider the theoretical foundations, algorithmic approaches, and practical implications when conducting studies within the field of granular computing. (NA?)

  • Carefully evaluate the trade-offs between the increased accuracy and reliability of Large Eddy Simulation (LES) and the higher computational costs and complexity when choosing between LES and Reynolds-averaged Navier-Stokes (RANS) simulations for various building simulation applications. (NA?)

  • Consider using the open-source toolkit Matminer to streamline data collection, feature extraction, and visualization processes in materials data mining projects, thereby improving efficiency and reproducibility. (NA?)

  • Carefully consider the unique characteristics of IoT data, such as volume, variety, velocity, and veracity, when choosing and implementing machine learning algorithms for effective data analysis and decision-making. (NA?)

  • Consider the full spectrum of smart manufacturing systems for Industry 4.0, covering areas like smart design, smart machining, smart monitoring, smart control, smart scheduling, and industrial applications, while taking into account the dimensions of sensor and actuator deployment, data collection, data analysis, and decision making. (NA?)

  • Consider utilizing deep learning techniques for analyzing electronic health records (EHR) data, as these methods have demonstrated superior performance compared to traditional machine learning approaches and require less time-consuming preprocessing and feature engineering. (NA?)

  • Carefully consider the tradeoff between the cost of remote-controlled switches (RCSs) and the potential reliability benefits when developing approaches for optimal RCS allocation in distribution systems. (NA?)

  • Consider employing deep neural networks, specifically fully convolutional networks (FCNs), when developing automated methods for analyzing cardiovascular magnetic resonance (CMR) images. These networks have the ability to accurately segment various regions of interest in the heart, such as the left ventricle (LV) and right ventricle (RV), while maintaining a level of precision comparable to that of human experts. Furthermore, they can do so rapidly, (NA?)

  • Adopt a combination of automation, high-throughput computing, and machine learning to accelerate materials development by 10x or more, thereby bridging the mismatch in time constants between materials research and market demands. (NA?)

  • Leverage the power of ChatGPT, a generative AI technology, in tandem with the Room2Educ8 framework, a design thinking-based methodology, to efficiently and effectively create high-quality educational escape rooms tailored to specific learning contexts. (NA?)

  • Utilize a combined approach of machine learning algorithms and high-throughput density functional theory calculations to effectively predict Debye temperature and band gap, which are proxies for photoluminescence quantum yield and structural rigidity, respectively. This integrated strategy enables the rapid screening of vast numbers of potential inorganic phosphor hosts, leading to the discovery of novel, highly efficient, and thermally stable materials. (NA?)

  • Utilize machine learning and artificial intelligence tools to enhance the efficiency and effectiveness of materials discovery for clean energy technologies. (NA?)

  • Consider incorporating a Graph-Structured Cache (GSC) into your models when dealing with open vocabularies in source code, as it significantly improves performance on code completion and variable naming tasks. (NA?)

  • Utilise a combination of advanced techniques including bot detection, multi-language sentiment analysis, network partitioning, and semantic network analysis to accurately identify and analyse the impact of bots on online social systems during politically charged events. (NA?)

  • Consider utilizing the cross ratio (CR) of Sentinel-1 VH/VV backscatter as it is largely effective in estimating Vegetation Water Content (VWC) across various crops and environmental conditions. (NA?)

  • Consider utilizing Random Forests (RF) instead of traditional multivariate regression techniques for improved accuracy in gap-filling and disaggregation of livestock data. (NA?)

  • Carefully consider the appropriate machine learning model and algorithm for your specific agricultural problem, taking into account factors such as data availability, desired outcome, and computational resources. (NA?)

  • Carefully select the most appropriate machine learning method for your specific building energy analysis, estimation, and benchmarking needs, taking into consideration factors such as data availability, model complexity, and desired level of accuracy. (NA?)

  • Consider employing machine learning (ML) methods for flood prediction, as they offer superior performance and cost-effectiveness compared to traditional physical models, while being able to accurately capture the complex mathematical expressions of physical processes related to floods. (NA?)

  • Investigate prompt engineering and iterative processes to understand how AI tools can be effectively and ethically incorporated into art and design education, thereby enhancing creative exploration, refining ideas, and improving the overall educational experience for students. (NA?)

  • Adopt a mixed-methods approach combining literature reviews, tool reviews, and expert interviews to understand the principles, components, roles, and architecture of Machine Learning Operations (MLOps) and ultimately contribute to a common understanding of the term and related concepts. (NA?)

  • Consider combining biochemical screening, network modeling, and machine learning to create a white-box machine learning approach for revealing drug mechanisms of action, which can help overcome the limitations of black-box machine learning and provide more interpretable and actionable insights. (NA?)

  • Consider using graph networks as a versatile machine learning framework for accurately predicting properties in both molecules and crystals, as demonstrated by the superior performance of the MEGNet models compared to prior ML models like SchNet in various tests. (NA?)

  • Consider employing hybrid machine learning models when working with energy systems, as they offer improved accuracy, robustness, precision, and generalization abilities, particularly in the context of renewable energy systems like solar, wind, and biofuels. (NA?)

  • Carefully select the appropriate machine learning algorithm based on the specific requirements of your smart transportation application, considering factors such as computational intensity, speed, and accuracy. (NA?)

  • Consider utilizing deep learning techniques when dealing with large amounts of heterogenous data in mobile and wireless networking contexts, due to its ability to effectively process complex correlations and perform hierarchical feature extraction. (NA?)

  • Consider employing transfer learning strategies when working with deep learning models across different datasets, as it allows for efficient adaptation to new data sources while maintaining high prediction accuracy. (NA?)

  • Consider utilizing quantum feature spaces for machine learning tasks, specifically through the development of quantum variational classifiers or quantum kernel estimators, to take advantage of the large dimensionality of quantum Hilbert space and potentially achieve significant computational speedups. (NA?)

  • Combine data generation with data-driven modeling to predict the behavior of complex and variable systems, such as lithium-ion batteries, using early-cycle data yet to exhibit capacity degradation. (NA?)

  • Integrate guideline-development efforts with substantive ethical analysis and adequate implementation strategies when investigating the global landscape of ethical AI principles and guidelines. (NA?)

  • Integrate machine learning techniques into your directed evolution workflow to significantly decrease experimental effort and enhance the efficiency of exploring the sequence space encoded by mutating multiple positions simultaneously. (NA?)

  • Carefully select and extract relevant features from the vast amounts of data generated by wind turbines, and apply suitable machine learning models to accurately detect and predict faults in order to reduce maintenance costs and improve overall efficiency. (NA?)

  • Consider incorporating machine learning techniques into your work, as these methods have shown great promise across various scientific disciplines, including physics, chemistry, and materials science. (NA?)

  • Consider developing specialised datasets like CropDeep for precision agriculture applications, as these datasets can significantly enhance the performance of deep-learning-based classification and detection models. (NA?)

  • Utilise machine learning techniques to streamline various aspects of the systematic review process, such as search, screening, and data extraction, while keeping humans “in-the-loop” to ensure accuracy and reliability. (NA?)

  • Consider using prompts derived from specific legal reasoning techniques, such as IRAC (Issue, Rule, Application, Conclusion), to achieve optimal results in legal reasoning tasks. (NA?)

  • Focus on utilising digital twins to enhance sustainable intelligent manufacturing by incorporating digital twin technology into the entire product lifecycle, from design to post-service, thus creating a more efficient, effective, and eco-friendly manufacturing process. (NA?)

  • Carefully consider the suitability of machine learning algorithms for specific fluid mechanics problems, taking into account factors such as interpretability, generalizability, cross-validation, and the presence of known physics in the system. (NA?)

  • Consider utilising machine learning (ML) techniques in additive manufacturing (AM) to improve the overall design and manufacturing workflow, specifically in areas like design for 3D printing, material tuning, process optimization, in-situ monitoring, cloud services, and cybersecurity. (NA?)

  • Carefully consider the selection of appropriate machine learning algorithms, predictor variables, and forcing datasets when developing upscaling models for estimating global carbon fluxes, as these choices significantly influence the accuracy and reliability of the resulting estimates. (NA?)

  • Carefully consider data selection, representation, model choice, regularization techniques, performance measurement, and result interpretation when applying machine learning to study complex systems such as metal-organic frameworks (MOFs) in porous materials. (NA?)

  • Consider applying machine learning techniques, particularly deep learning algorithms, to optimize the performance of future wireless networks across various layers, from physical to application layers, due to your ability to handle large volumes of data and adapt to changing environments. (NA?)

  • Aim to create explainable AI (XAI) systems that allow users to understand and explain the models predictions, while also developing standards for AI applications in areas such as concepts and terminology, data principles, sample size estimation, metrics, performance testing and methodology, risk management, and value and trustworthiness.’ (NA?)

  • Develop hybrid brain-machine interfaces (HBMIs) that utilize both invasive and non-invasive methods for sampling brain activity, enabling real-time interaction between the brain and external devices, thereby potentially revolutionizing the way individuals perceive, think, and act. (NA?)

  • Conduct a comprehensive literature review on machine learning approaches for securing IoT systems, focusing on the importance of security in terms of different types of possible attacks, and presenting potential ML-based solutions for IoT security along with future challenges. (NA?)

  • Consider the potential negative impacts of AI on Sustainable Development Goals (SDGs) alongside the positive impacts, and strive to minimize these negative impacts through careful design and deployment of AI technologies. (NA?)

  • Utilize the CONSORT-AI checklist to ensure comprehensive and transparent reporting of AI intervention trials, thereby promoting critical appraisal and evidence synthesis. (NA?)

  • Combine both self-reported symptoms and sensor metrics from smartwatches and activity trackers to achieve better accuracy in distinguishing between symptomatic individuals with and without a diagnosis of COVID-19. (NA?)

  • Consider utilizing crowdsourced data collection frameworks, such as mobile apps, to gather large-scale, diverse datasets of COVID-19 related sounds, which can potentially aid in developing accurate diagnostic tools for distinguishing COVID-19 users from healthy ones. (NA?)

  • Conduct a comprehensive and systematic review of Artificial Intelligence (AI) techniques used in energy demand-side response (DR) applications, considering the advantages and drawbacks of each approach in specific domains, and identify potential research gaps and future research paths in this rapidly evolving field. (NA?)

  • Consider utilizing chest x-ray (CXR) imaging as a supplementary tool for COVID-19 screening, particularly in resource-limited settings, due to its rapid triage capabilities, wide availability, and portability, which allows for safer and quicker diagnosis compared to other methods like CT imaging or RT-PCR tests. (NA?)

  • Consider employing deep learning (DL) techniques in structural health monitoring (SHM) applications, as they allow for automatic feature extraction and hierarchical representation mechanisms from raw input data, making them suitable for large-scale structures and efficient for vision- and vibration-based SHM. (NA?)

  • Consider the implications of industrys increasing dominance in Artificial Intelligence (AI) research, particularly in terms of talent acquisition, computational power, and large dataset availability, which could potentially lead to a lack of public interest alternatives for crucial AI tools.’ (NA?)

  • Conduct a systematic literature review using multiple sources and analytical tools to better understand the current state and potential directions of artificial intelligence and machine learning applications in smart production. (NA?)

  • Consider using deep learning (DL) algorithms, specifically convolutional neural networks (CNNs), recurrent neural networks (RNNs), and autoencoders, when analyzing biological data such as images, signals, and sequences, as these methods have demonstrated success in accurately recognizing, classifying, and predicting patterns in large and complex datasets. (NA?)

  • Utilize an empirical Bayesian algorithm called SpEM to iteratively eliminate spammers and estimate consensus labels based solely on good annotators, resulting in improved accuracy and reduced costs. (NA?)

  • Conduct a comprehensive review of the application of machine learning techniques for big data analysis in the healthcare sector, focusing on the strengths and weaknesses of existing methods and highlighting various research challenges to guide future work. (NA?)

  • Consider combining machine learning techniques with traditional numerical methods to achieve significant improvements in computational efficiency and accuracy for simulations of complex physical systems governed by nonlinear partial differential equations. (NA?)

  • Use a combination of quantitative methods, network analysis, and machine learning techniques to analyze the complex dynamics of the NFT market, taking into consideration various factors such as temporal patterns, network structures, and visual features. (NA?)

  • Adopt a multi-disciplinary approach to machine learning in agriculture, focusing on diverse applications such as crop management, water management, soil management, and livestock management, and leveraging a wide array of machine learning algorithms, particularly Artificial Neural Networks, to effectively address the challenges faced by the agricultural industry. (NA?)

  • Exercise caution while utilizing Large Language Models (LLMs) like ChatGPT due to your tendency to generate erroneous and misleading information, particularly for highly specialized or technical subjects. (NA?)

  • Carefully consider the ethics and implications of integrating AI chatbots like ChatGPT into higher education, especially regarding assessment, learning, and teaching practices. (NA?)

  • Employ a combination of qualitative and quantitative methods, including longitudinal studies and experimental designs, to thoroughly examine the impact of AI technologies like ChatGPT on education, considering factors such as ethics, digital literacy, and the evolving role of educators. (NA?)

  • Consider the potential for text-to-image generation models to amplify and perpetuate stereotypes, regardless of whether prompts explicitly mention identity and demographic language or avoid such language, and despite mitigation strategies such as user attempts to counter stereotypes or institutional efforts to implement system guardrails. (NA?)

  • Consider using Latent Dirichlet Allocation (LDA) topic modeling algorithm to analyze social media data like tweets, in order to extract meaningful insights and patterns related to emerging technologies like ChatGPT. (NA?)

  • Consider exploring the potential of Generative Pre-Trained Transformer (GPT) language models like ChatGPT for automating construction scheduling tasks, while acknowledging the limitations and need for further development in the technology. (NA?)

  • Critically analyze ChatGPT output, compare it to established research, and adapt it to your specific teaching contexts, while being mindful of ethical concerns such as environmental impact, content moderation, and copyright infringement. (NA?)

  • Adopt a critical approach when evaluating the output of language models like ChatGPT, recognizing your limitations in terms of bias, accuracy, and lack of independent scientific reasoning, while exploring technologies such as blockchain to enhance the security and originality of scientific projects. (NA?)

  • Focus on developing effective prompt engineering skills to maximize the potential of generative AI in education, while addressing challenges related to data privacy, security, and ethical considerations. (NA?)

  • Focus on developing a principled graph structure learning (GSL) approach for reducing edge noise in social graphs, particularly when dealing with large-scale social graphs where traditional similarity-guided GSL methods become computationally expensive. (NA?)

  • Embrace open, reproducible, and replicable research practices, utilize shared research resources, and engage in collaborative efforts to create sustainable research environments. (NA?)

  • Explore using ChatGPT, a large language model, to make complex medical information like CPR guidelines more accessible and understandable to the general public, despite potential risks associated with relying on AI-generated information. (NA?)

  • Consider the “jagged technological frontier” when studying the impact of AI on knowledge work, recognizing that AI capabilities vary across tasks and can lead to significant increases in quality and productivity for certain tasks, while potentially decreasing performance for others. (NA?)

  • Carefully evaluate the accuracy and reliability of AI-generated content, particularly regarding references and synthesis of complex information, and consider combining AI assistance with human oversight to optimize the efficiency and quality of scientific writing. (NA?)

  • Utilise a combination of multi-turn conversations, domain-specific training, and user feedback to effectively leverage large language models (LLMs) in the development of cognitive simulations. (NA?)

Natural Language Processing (Nlp)

  • Consider employing weak supervision, transfer learning, and prompt engineering techniques to maximize the performance of language models for social data science tasks, even with limited labeled training data. (Castro-Gonzalez et al. 2024)

  • Explore the use of large language models (LLMs) like GPT-3, Alpaca, and FLAN-T5 for early depression detection, while considering potential issues such as data bias and ethical concerns prior to integrating these technologies into clinical practice. (Chowdhury et al. 2024)

  • Employ a lightweight large language model (LLM) like PokerGPT to efficiently and effectively solve imperfect information games (IIGs) like Texas Holdem Poker, as it offers numerous benefits such as lower computational costs, faster response times, and greater adaptability to multi-player environments.’ (Chenghao Huang et al. 2024)

  • Create a high-quality dataset with expert-annotated images and establish a comprehensive evaluation framework consisting of four dimensions - perception, empathy, assessment, and interpretation - to thoroughly assess the performance of multimodal large language models in image aesthetics perception. (Yipo Huang et al. 2024)

  • Focus on understanding the origins of hallucinations in large language models (LLMs) by investigating the effects of pre-training, fine-tuning, prompt design, and inference methods on hallucination rates, and subsequently develop targeted mitigation strategies accordingly. (Junyi Li et al. 2024)

  • Utilize an unsupervised approach with minimal human intervention to achieve neuron-level explainability for language models, specifically by taking advantage of open-source generative language models to generate meaningful descriptors and using classical NLP techniques such as clustering to obtain candidate descriptors, which can then be automatically assigned to neurons of a widely used transformer-based model to better understand your behavior and contribution to various downstream tasks. (Mondal et al. 2024)

  • Consider adopting hybrid approaches that integrate machine learning and symbolic methods to address the limitations of each method separately, thereby producing more accurate and reliable results in natural language processing tasks. (Panchendrarajan and Zubiaga 2024)

  • Utilise a combination of in-context learning and demonstrations to improve translation performance in machine translation tasks. (Pourkamali and Sharifi 2024)

  • Use a combination of convolutional neural networks (CNN) and bidirectional long short-term memory (BiLSTM) architectures, incorporating pretrained GloVe and FastText embeddings, for effective gendered abuse detection in Indic languages. (Vaidya et al. 2024)

  • Utilize the APT (Adaptive Pruning and Tuning) methodology to optimize the efficiency of training and inference processes in large language models, thereby achieving significant reductions in memory usage and computational time without compromising model performance. (Bowen Zhao, Hajishirzi, and Cao 2024)

  • Focus on responsible AI development to ensure the safe integration of generative AI technologies, such as ChatGPT, into education systems, while acknowledging the potential for profound changes in teaching and learning methods. (Bozkurt and Sharma 2023)

  • Explore the perspectives, experiences, and strategies employed by early adopters of novel pre-trained models like Llama 2 to understand your strengths, weaknesses, and potential areas for improvement, ultimately guiding the effective application and development of AI solutions. (Roumeliotis, Tselikas, and Nasiopoulos 2023)

  • Capitalize on the numerous opportunities afforded by ChatGPT in the research lifecycle, ranging from idea generation to manuscript preparation, while remaining cognizant of the challenges such as AI authorship, nonexistent references, unintentional plagiarism, biases, and inaccuracies, ensuring that the ultimate decision-making power rests with humans. (“ChatGPT for Research and Publication: Opportunities and Challenges” 2023)

  • Adopt a multi-modal summarization (MMS) framework that integrates natural language processing (NLP), speech processing, and optical character recognition (OCR) watermarking techniques to analyze elaborate information across various modes and generate accurate, concise, and informative summaries. (Jangra et al. 2023)

  • Conduct a comprehensive survey with students and teachers to determine how ChatGPT supports programming learning and teaching, and investigate the opportunities and possible threats of using ChatGPT in educational settings, particularly in programming education. (Md. M. Rahman and Watanobe 2023)

  • Carefully balance the growth of model size and data size in order to optimize performance, rather than simply increasing model size alone. (Anil et al. 2023)

  • Consider users broader context and goals beyond prompt/chain iteration itself, drawing inspiration from past frameworks for data processing.’ (Arawjo et al. 2023)

  • Utilise the JEEBench dataset to challenge and advance the problem-solving abilities of large language models (LLMs), focusing on long-horizon reasoning and deep in-domain knowledge. (D. Arora and Singh 2023)

  • Focus on developing large language models (LLMs) with diverse training data, advanced skillsets, and specialized models for specific tasks, while addressing issues of reproducibility, steerability, and accessibility through open-sourcing and collaboration. (J. Bai et al. 2023)

  • Consider employing a comprehensive evaluation framework for interactive large language models (LLMs) such as ChatGPT, encompassing multiple dimensions like multitasking, multilingualism, and multimodality, using diverse publicly available datasets spanning various NLP applications. (Y. Bang et al. 2023)

  • Utilize textual abstractions of process mining artifacts, such as event logs and process models, to enable effective communication with large language models (LLMs) like GPT-4, thereby facilitating accurate interpretation and analysis of complex process structures. (Berti, Schuster, and Aalst 2023)

  • Focus on improving large language models ability to identify and leverage necessary commonsense knowledge for answering specific questions, particularly in areas such as social, event, and temporal domains.’ (Bian et al. 2023)

  • Utilise scaling laws to predict the behaviour of large language models before they are trained, taking into consideration the unusual scaling patterns and emergence of semi-emergent behaviour as model scale increases. (Biderman et al. 2023)

  • Carefully select your prompt types and settings when working with GPT-3, as different combinations can lead to significant variations in performance. (Blair-Stanek, Holzenberger, and Durme 2023)

  • Combine multiple large language models to create an intelligent agent system capable of autonomous design, planning, and execution of scientific experiments, while considering safety implications and proposing measures to prevent misuse. (Boiko, MacKnight, and Gomes 2023)

  • Carefully consider the limitations and potential risks associated with large language models, particularly regarding your scalability, unpredictable emergence of behaviors, lack of reliable control techniques, and difficulty interpreting your inner workings. (Bowman 2023)

  • Integrate expert-designed tools into large language models (LLMs) to overcome your limitations in handling chemistry-related problems, ultimately enabling LLMs to become powerful assistants in various chemical tasks. (Bran et al. 2023)

  • Explore the potential of prompt engineering in business process management (BPM) to effectively utilize limited data volumes, facilitate natural language-based interaction, optimize input via prompt templates, overcome task-specificity, improve computational efficiency, and increase explainability. (Busch et al. 2023)

  • Utilize a pre-trained BERT model specifically designed for the Spanish language, rather than a multi-lingual model, to achieve superior results in various Spanish language tasks. (Cañete et al. 2023)

  • Utilize a combination of ensemble methods based on mixed statistical and comparative features along with neural information retrieval approaches to effectively rank and retrieve comparative arguments from large datasets. (Chekalina and Panchenko 2023)

  • Consider applying computational psychiatry techniques to study the behaviour of large language models, as demonstrated through the authors investigation of GPT-3.5’s response to anxiety questionnaires and emotion-inducing prompts.’ (Coda-Forno et al. 2023)

  • Consider utilising ChatGPT for data preparation tasks, as demonstrated through various examples where the tool successfully created data frames, pivoted tables, derived new columns, and calculated correlations, all without requiring any manual coding. (Peng Ding 2023)

  • Consider using embodied language models to integrate real-world continuous sensor modalities into language models, thus facilitating grounded inferences for sequential decision making in the real world. (Driess et al. 2023)

  • Develop robust success detectors that leverage large, pretrained vision-language models (such as Flamingo) and human reward annotations to improve generalization abilities across different domains and reduce the need for extensive labor-intensive annotations. (Yuqing Du, Konyushkova, et al. 2023)

  • Carefully consider the potential privacy risks associated with using prompts in large language models, and explore techniques like PromptDPSGD and PromptPATE to ensure differential privacy while maintaining high utility. (H. Duan et al. 2023)

  • Consider implementing a prototype called CoPrompt’, which offers mechanisms such as referring, requesting, sharing, and linking to support collaborative prompt engineering in natural language programming.’ (F. L. Feng et al. 2023)

  • Consider developing dynamic large language models on blockchains to enable continuous learning from user input, thereby providing a novel approach to developing large language models and potentially paving the way for next-generation artificial intelligence systems. (Yuanhao Gong 2023)

  • Utilise the Eigenvalue-corrected Kronecker-Factured Approximate Curvature (EK-FAC) parameterisation to approximate the Hessian in order to overcome the computational bottlenecks associated with calculating inverse-Hessian-vector products (IIVP) in large transformer language models. (Grosse et al. 2023)

  • Adopt a collaborative approach when building legal reasoning benchmarks for large language models, actively involving legal professionals in the creation of evaluation tasks to ensure accurate representation of various types of legal reasoning. (Guha et al. 2023)

  • Combine large language models (LLMs) with evolutionary algorithms (EAs) to create powerful prompt optimizers, allowing for more efficient and accurate optimization of prompts for various language understanding and generation tasks. (Q. Guo et al. 2023)

  • Consider developing and utilizing multimodal large language models (MLLMs) like Kosmos-1, which can perceive general modalities, follow instructions, learn in context, and generate outputs, enabling improved performance in various language, perception-language, and vision tasks. (Shaohan Huang et al. 2023)

  • Leverage large language models (LLMs) to develop conversational agents capable of directly extracting answers from event data, thereby reducing the need for interaction with multiple stakeholders. (Jessen, Sroka, and Fahland 2023)

  • Leverage the asymmetry in task difficulty between generating structured outputs and producing plausible input text for those outputs, allowing them to create large-scale, high-quality synthetic data for complex tasks like closed information extraction. (Josifoski et al. 2023)

  • Carefully consider the composition of pre-training corpora, including the balance of domain mixtures and fine-tuning task mixtures, to ensure optimal performance and minimize privacy risks. (Kaddour et al. 2023)

  • Consider utilizing large language models, specifically GPT-4, for evaluating and potentially solving complex problems in fields such as law, given its impressive performance on the Uniform Bar Examination, which includes multiple-choice, essay, and performance test components. (D. M. Katz et al. 2023)

  • Carefully distinguish between different types of causal reasoning tasks, such as causal discovery, counterfactual reasoning, and actual causality, and recognize the unique challenges and limitations of each task when working with large language models (LLMs) to ensure accurate and reliable results. (Kıcıman et al. 2023)

  • Employ stylometric features alongside pre-trained language models to enhance the accuracy of detecting AI-generated tweets in a given users Twitter timeline.’ (Kumarage et al. 2023)

  • Employ a novel prompt tuning method within a few-shot learning context to recognize keywords and parameters in log messages, allowing for accurate and efficient log parsing. (V.-H. Le and Zhang 2023)

  • Carefully curate and refine your training datasets, paying close attention to data quality, redundancy reduction, and contamination control, in order to optimize model performance and reduce training costs. (A. N. Lee, Hunter, and Ruiz 2023)

  • Focus on developing a web-based code generation platform using ChatGPT and Prompt Engineering to optimize the performance of the model in generating code, leading to improvements in metrics such as EM, BLEU, CodeBLEU, and Pass@1. (Youjia Li, Shi, and Zhang 2023)

  • Focus on decoupling schema linking and skeleton parsing in Text-to-SQL tasks through a ranking-enhanced encoding and skeleton-aware decoding framework, leading to increased efficiency and effectiveness in generating accurate SQL queries. (Haoyang Li et al. 2023)

  • Combine the strengths of large language models (LLMs) and classical planners through the LLM+P framework, allowing LLMs to effectively solve long-horizon robot planning problems by converting natural language descriptions into PDDL format, leveraging classical planners for quick identification of correct or optimal plans, and translating the found solution back into natural language. (Bo Liu et al. 2023)

  • Consider using EvalPlus, a code synthesis evaluation framework that generates diverse test inputs using both LLM- and mutation-based strategies, to rigorously benchmark the functional correctness of LLM-synthesized code and avoid potential issues caused by insufficient testing and imprecise problem descriptions. (Jiawei Liu et al. 2023)

  • Leverage the capabilities of Large Language Models (LLMs) through prompt engineering when developing news recommendation systems, as demonstrated by the RecPrompt framework, which incorporates a prompt optimizer that applies an iterative bootstrapping process to enhance the LLM-based recommenders ability to align news content with user preferences and interests more effectively.’ (D. Liu et al. 2023)

  • Consider employing mixed prompt settings during training, as it leads to significant improvements in both zero-shot and few-shot performance, regardless of model size. (Longpre et al. 2023)

  • Consider incorporating advanced language models like ChatGPT into your investment decision-making processes to achieve more accurate predictions and improve the performance of quantitative trading strategies. (Lopez-Lira and Tang 2023)

  • Consider developing a plug-and-play compositional reasoning framework, called Chameleon, that enables large language models (LLMs) to synthesize programs and compose various tools for a wide range of tasks, thereby addressing inherent limitations of LLMs and creating a versatile and adaptable AI system. (P. Lu et al. 2023)

  • Consider implementing an iterative refinement approach called Self-Refine, which involves generating an initial output using a large language model (LLM), providing feedback for its output, and using it to refine itself, iteratively. This method does not require any supervised training data, additional training, or reinforcement learning, and instead uses a single LLM as the generator, refiner, and feedback provider. (Madaan et al. 2023)

  • Consider using a combination of objective and subjective questions in order to achieve a comprehensive and balanced assessment of large language models (LLMs) in the AIOps domain. (Yukai Miao et al. 2023)

  • Utilise a combination of transformer-based machine learning models and explainable artificial intelligence frameworks, such as SHAP, to effectively train models to distinguish between human and ChatGPT-generated text, and to gain valuable insights into the reasoning behind these models decisions.’ (Mitrović, Andreoletti, and Ayoub 2023)

  • Optimize the use of prompts in large language models (LLMs) to effectively guide your application in clinical decision-making, especially focusing on the integration of domain-specific knowledge derived from interpretable machine learning models. (Nazary, Deldjoo, and Noia 2023)

  • Carefully evaluate the performance of large language models (LLMs) on medical competency examinations and benchmark datasets to understand your capabilities and limitations in the medical domain. (Nori et al. 2023)

  • Carefully evaluate the tradeoffs between conversational AI models and traditional question-answering systems for knowledge graphs, taking into account factors such as accuracy, robustness, explainability, and the ability to incorporate recent information, in order to determine the optimal approach for your specific application. (Omar et al. 2023)

  • Consider fine-tuning pre-trained language models with datasets containing both code review and subsequent code changes to significantly enhance the performance of automated program repair systems. (R. Paul et al. 2023)

  • Utilize GPT-4 to generate instruction-following data for large language models (LLMs) finetuning, leading to improved zero-shot performance on new tasks. (B. Peng et al. 2023)

  • Employ “conversational large language models” (CLLMs) to efficiently and accurately extract data from research papers, using a carefully designed set of prompts to guide the CLLM towards identifying relevant sentences, extracting data, and ensuring its correctness through follow-up questions. (Polak and Morgan 2023)

  • Consider decomposing complex tasks like text-to-SQL into smaller sub-tasks, allowing large language models to effectively handle these tasks by leveraging your strengths in handling individual sub-problems. (Pourreza and Rafiei 2023)

  • Utilize the ProTeGi algorithm for automatic prompt optimization, which combines gradient descent principles with beam search and bandit selection methods to efficiently optimize prompts for large language models, resulting in improved performance and reduced reliance on manual efforts. (Pryzant et al. 2023)

  • Consider using large language models (LLMs) like ChatGPT to enhance molecular property prediction tasks by generating more accurate and informative textual descriptions of molecules, leading to improved performance in downstream tasks. (C. Qian et al. 2023)

  • Employ a general tool-use framework like ToolLLM, which includes data construction, model training, and evaluation, to enable open-source large language models to effectively interact with various tools (APIs) and accomplish complex tasks. (Yujia Qin, Liang, et al. 2023)

  • Consider the effects of linguistic nuances in prompts, particularly readability, formality, and concreteness, on reducing hallucinations in large language models. (Rawte et al. 2023)

  • Utilise the Progressive Prompts methodology for continual learning in language models, as it enables forward transfer and mitigates catastrophic forgetting without requiring data replay or numerous task-specific parameters. (Razdaibiedina et al. 2023)

  • Focus on developing scalable model architectures and efficient distributed training systems to maximize training throughput and optimize performance in large language models. (X. Ren et al. 2023)

  • Be cautious about overestimating the effectiveness of current AI-text detectors due to your susceptibility to evasion tactics such as paraphrasing attacks and the inherent difficulty in distinguishing between human-generated and AI-generated text as language models improve. (Sadasivan et al. 2023)

  • Use the TELeR taxonomy to classify and standardize LLM prompts for complex tasks, enabling better comparisons and evaluations across studies. (Santu and Feng 2023)

  • Carefully consider the choice of metric when studying emergent abilities in large language models, as nonlinear or discontinuous metrics can create the appearance of sudden and unpredictable changes in model performance, while linear or continuous metrics result in smoother, more predictable improvements. (Schaeffer, Miranda, and Koyejo 2023)

  • Investigate the interaction between humans and large language models (LLMs) in competitive situations, specifically focusing on price negotiations, to identify negotiation strategies, outcomes, and potential threats posed by LLMs, such as your susceptibility to prompt hacking and reasoning deficiencies. (Schneider, Haag, and Kruse 2023)

  • Consider leveraging pre-trained Large Language Models (LLMs) for efficient and effective extraction of astronomical knowledge entities from astrophysical journal articles, while taking into account factors influencing LLM performance such as token limitations, prompt design, and choice of LLM. (W. Shao et al. 2023)

  • Carefully curate a diverse set of questions across multiple domains and evaluate the performance of large language models like ChatGPT using both correctness and unanswerable question identification measures, considering the potential impact of system roles and adversarial examples on reliability. (Xinyue Shen et al. 2023)

  • Leverage the power of large language models like GPT to build autonomous edge AI systems that can automatically organize, adapt, and optimize themselves to meet users diverse requirements, while ensuring privacy and low latency.’ (Yifei Shen et al. 2023)

  • Carefully examine the tradeoffs between fine-tuning and prompt engineering when working with large language models like GPT-4 in automated software engineering tasks, considering aspects such as performance, cost, and ease of implementation. (J. Shin et al. 2023)

  • Focus on developing automated strategies for generating and selecting optimal chain-of-thought prompts, rather than relying solely on human-engineered approaches, to enhance the reasoning capabilities of large language models. (Shum, Diao, and Zhang 2023)

  • Pay close attention to the selection of sample questions when using few-shot prompting, as poor choices can introduce noise detrimental to the task at hand. (Linxin Song et al. 2023)

  • Pay attention to the selection of instructional techniques for large language models (LLMs) in passage re-ranking tasks, as it can impact the overall effectiveness of the model. (W. Sun et al. 2023)

  • Use a novel training paradigm involving the generation of high-quality synthetic data with labels using large language models (LLMs) and subsequently fine-tuning a local model for downstream tasks, resulting in significant improvements in performance and reduced privacy concerns. (R. Tang et al. 2023)

  • Carefully consider the potential impact of data leakage when evaluating the performance of large language models (LLMs) in software engineering tasks, particularly when using publicly available benchmark data sets that might have been included in the training corpus of the LLM being tested. (H. Tian et al. 2023)

  • Carefully validate the performance of Large Language Models (LLMs) on specific tasks, ensuring that they accurately measure intended concepts without introducing problematic biases, despite your impressive capabilities in areas such as zero-shot learning and classification. (Törnberg 2023)

  • Prioritize training larger language models on more tokens rather than focusing solely on increasing model size, as doing so can result in improved performance at lower computational costs. (Touvron et al. 2023)

  • Focus on developing and releasing large language models (LLMs) that are pretrained on diverse datasets and then fine-tuned for specific use cases, such as dialogue, to achieve optimal performance and safety. (Touvron et al. 2023)

  • Develop a benchmark suite based on the International Planning Competition to systematically evaluate the planning capabilities of large language models (LLMs) in autonomous, heuristic, and human-in-the-loop modes. (Valmeekam et al. 2023)

  • Consider implementing prompt injection as a quality assurance tool to manipulate and detect LLM-generated responses in crowdsourced surveys, thus ensuring the integrity of collected data. (Chaofan Wang et al. 2023)

  • Consider using cross-lingual summarization (CLS) as a testbed for evaluating the capabilities of large language models (LLMs) in handling complex tasks requiring simultaneous translation and summarization. (Lei Wang and Lim 2023)

  • Consider utilizing ChatGPT for systematic review Boolean query construction and refinement, while acknowledging its limitations and caveats. (Jun Wang et al. 2023)

  • Prioritize safety-capability parity, meaning that safety mechanisms should be as sophisticated as the underlying model, and recognize that simply scaling up the model might not resolve safety failure modes. (A. Wei, Haghtalab, and Steinhardt 2023)

  • Utilize a catalog of prompt patterns to systematically engineer prompts for large language models (LLMs) in order to address various issues encountered during conversations and output generation for automating software development tasks. (J. White, Fu, et al. 2023)

  • Develop and utilize prompt patterns for software engineering tasks, which are reusable prompt designs that solve common problems in LLM interaction, improving software quality attributes such as modularity or reusability. (J. White, Hays, et al. 2023)

  • Consider implementing the StreamingLLM framework, which allows large language models trained with a finite attention window to generalize to infinite sequence length without fine-tuning, thus improving your ability to handle long sequences in streaming applications. (G. Xiao et al. 2023)

  • Consider utilizing large language models (LLMs) in communication games, specifically proposing a tuning-free framework that relies on retrieval and reflection on past communications and experiences for improvement, as evidenced by successful application in the Werewolf game. (Yuzhuang Xu et al. 2023)

  • Investigate how large language models (LLMs) can be integrated with knowledge graph completion (KGC) tasks, focusing on the potential benefits of using LLMs to enhance textual descriptions and reduce hallucination issues through contextualized prompts. (Rui Yang, Fang, and Zhou 2023)

  • Consider using a Generative Multimodal Prompt (GMP) model for few-shot Multimodal Aspect-Based Sentiment Analysis (MABSA), which includes the Multimodal Encoder (ME) module and the N-Stream Decoders (NSD) module, to effectively handle the challenges posed by the unknown and varying number of aspect items in each sample. (Xiaocui Yang et al. 2023)

  • Carefully consider the unique challenges presented by prompt watermarking, such as low entropy prompts and difficulties in verifying watermarks in sequence classification tasks, and develop tailor-made solutions for these issues. (Hongwei Yao et al. 2023)

  • Develop a benchmark construction protocol that ensures clear differentiation and challenging distribution shifts to accurately evaluate out-of-distribution (OOD) robustness in NLP models. (Lifan Yuan et al. 2023)

  • Carefully consider the impact of tokenization, pre-training, prompts, interpolation and extrapolation, scaling laws, chain-of-thought, and instruction control on the arithmetic ability of large language models. (Zheng Yuan, Yuan, Tan, et al. 2023)

  • Carefully assess the quality of reference summaries when evaluating large language models for news summarization, as low-quality references can lead to underestimation of human performance and negatively impact the performance of systems trained through finetuning or few-shot prompting. (Tianyi Zhang et al. 2023)

  • Consider developing a cross-lingual neural codec language model like VALL-E X for cross-lingual speech synthesis, as it allows for high-quality zero-shot cross-lingual speech synthesis while preserving the speakers voice, emotion, and acoustic environment.’ (Ziqiang Zhang et al. 2023)

  • Consider developing and releasing large pre-trained multilingual code generation models, such as CodeGeeX, to demonstrate consistent outperformance on code generation and translation tasks over multilingual baseline models of the same scale, while supporting diverse functions including code completion, generation, translation, and explanation. (Q. Zheng et al. 2023)

  • Consider implementing “Batch Calibration” (BC) as a simple, zero-shot, inference-only calibration method for in-context learning (ICL) that effectively mitigates bias from the batch and improves overall performance. (Han Zhou et al. 2023)

  • Integrate large language models (LLMs) into information retrieval (IR) systems to improve your performance, specifically by enhancing query rewriting, retrieval, reranking, and reading components. (Yutao Zhu et al. 2023)

  • Consider implementing SpikeGPT, a generative language model with binary, event-driven spiking activation units, to improve the energy efficiency of large language models while maintaining competitive performance. (R.-J. Zhu et al. 2023)

  • Leverage the power of ChatGPT to automate various aspects of your workflow, such as data cleaning and preprocessing, model training, and result interpretation, while remaining aware of potential biases and limitations in the models outputs.’ (Hassani and Silva 2023)

  • Consider using compact transformer models, such as DistilBERT, MobileBERT, and TinyBERT, for biomedical NLP tasks, as they offer significant improvements in efficiency and speed while still providing comparable performance to larger models. (Rohanian et al. 2023)

  • Consider using large language models like FinBERT for textual analysis in financial economics, as it outperforms traditional NLP algorithms and other deep learning methods in sentiment classification and ESG identification tasks, particularly when dealing with small training samples. (A. H. Huang, Wang, and Yang 2023)

  • Prioritize the development and application of unsupervised or semi-supervised learning techniques in natural language processing (NLP) research, as these methods can effectively learn from unannotated or partially annotated data, thereby maximizing the potential of the vast quantities of available unlabeled data on the internet. (Bharadiya 2023)

  • Develop customizable prompts for ChatGPT to enable efficient translation of natural-language instructions into executable robot actions, allowing for seamless integration with robot execution systems or visual recognition programs, adaptation to various environments, and creation of multi-step task plans while addressing the token limit issue. (Wake et al. 2023)

  • Carefully consider the ethical implications of using NLP models in education, especially regarding academic integrity and fairness, due to your potential to facilitate cheating and plagiarism. (Lesage et al. 2023)

  • Utilize machine learning techniques, specifically BERT-based models, to effectively distinguish between medical texts authored by humans and those produced by ChatGPT, thereby promoting responsible and ethical use of artificial intelligence in the generation of medical content. (W. Liao et al. 2023)

  • Utilize a combination of Hearst-like lexico-syntactic patterns and a comprehensive web corpus to effectively extract hypernymy relations across various domains. (Hubert et al. 2023)

  • Consider integrating large language models with computational interactive agents to achieve believable simulations of human behavior. (J. S. Park et al. 2023)

  • Consider utilizing large language models like ChatGPT for text annotation tasks instead of traditional methods like crowd-working platforms, as they offer higher accuracy, better intercoder agreement, and significantly reduced costs. (Gilardi, Alizadeh, and Kubli 2023)

  • Consider employing a “Prompt Generation” approach when dealing with Visual Word Sense Disambiguation tasks. This involves utilizing a large language model like ChatGPT to create tailored prompts for each sense of a word, thereby enhancing the correlation between textual and visual contexts within a multimodal model like CLIP. This strategy enhances the robustness of such models against contextual ambiguities, leading to improved performance in tasks such as image retrieval and ca (Ghahroodi et al. 2023)

  • Validate the use of generative AI models like ChatGPT in SPC through proper validation and combination with other methods to ensure accurate results. (Megahed et al. 2023)

  • Consider utilizing both word and network embeddings when attempting to predict Wikipedia infobox types, as this approach offers improved performance over traditional methods like TF-IDF. (Biswas et al. 2023)

  • Consider the potential impact of AI-generated text on the reliability and validity of your findings, particularly in areas where accurate representation of data is crucial, and explore strategies to mitigate any negative effects. (Brainard 2023)

  • Utilize ChatGPT, a powerful AI tool, to generate high-quality, accurate, and readable clinical letters for patients, thereby enhancing efficiency, consistency, and patient satisfaction, but they should also exercise caution and maintain a human-in-the-loop approach to verify the outputs due to potential risks related to misinterpretation or omission of crucial information. (S. R. Ali et al. 2023)

  • Carefully consider the potential benefits and limitations of using ChatGPT in healthcare education, research, and practice, while also taking into account ethical implications and the need for clear guidelines on its usage. (Sallam 2023)

  • Carefully encode and adjudicate ChatGPTs responses to multiple-choice questions, considering factors like accuracy, agreement with the answer key, and insight, to effectively evaluate its potential as a valuable study tool for premedical students.’ (Bommineni et al. 2023)

  • Carefully balance the representation of fundamental and specialized knowledge in AI training datasets to optimize performance in medical knowledge assessment tasks. (Teebagy et al. 2023)

  • Utilise the HaluEval benchmark to analyse the types of content that large language models (LLMs) like ChatGPT tend to hallucinate, and subsequently develop effective methods to mitigate this issue. (Junyi Li et al. 2023)

  • Consider the unique characteristics of poetry, including its figurative language and deeper meaning, when developing models for poem summarization, as traditional text summarization techniques may not effectively capture the essence of a poem. (Mahbub et al. 2023)

  • Leverage the strengths of GPT-4 in areas such as few-shot learning, understanding of chemical concepts, and ability to extract relevant variables from datasets, while acknowledging its current limitations in areas such as numerical recognition, handling of complex molecular structures, and inability to access academic literature. (A. D. White et al. 2022b)

  • Consider developing a multilingual sign language translation (MSLT) model instead of relying solely on bilingual sign language translation (BSLT) models, as the former allows for a more efficient and effective handling of multiple sign languages and spoken languages. (A. Yin et al. 2022)

  • Focus on developing a reinforcement learning-based method called Reinforce-Detoxify, which incorporates a novel reward model capable of detecting toxic content and mitigating unintended bias towards social identities in toxicity prediction, thereby improving the performance of language model detoxification. (Faal, Schmitt, and Yu 2022)

  • Consider the impact of temporal modeling, video-to-text multimodal fusion, masked modeling objectives, and joint training on images and videos when developing effective video-and-language pretraining models. (Bain et al. 2022)

  • Combine semantic and acoustic tokens in a hierarchical manner to achieve long-term consistency and high quality in audio generation. (Borsos et al. 2022)

  • Utilize AdaPrompt, an adaptive model training approach for prompt-based NLP, to enhance the performance of pretrained language models in zero-shot and few-shot scenarios by leveraging external data for continual pretraining and employing knowledge from Natural Language Inference models for deriving adaptive verbalizers. (Yulong Chen et al. 2022)

  • Focus on scaling up language models to improve your performance in few-shot learning tasks, as evidenced by the success of the 540-billion parameter Pathways Language Model (PaLM) in achieving state-of-the-art results on numerous language understanding and generation benchmarks. (Chowdhery et al. 2022)

  • Carefully consider the combination of various scoring methods when performing named entity linking tasks, particularly focusing on entity popularity, entity-content similarity, and entity-entity similarity, as demonstrated through the successful application of UIScore and UCSE in both Quotebank and AIDA-CoNLL datasets. (Čuljak et al. 2022)

  • Consider the importance of integrating multiple modes of communication, such as voice and visual cues, in conversational AI systems to create a more comprehensive and engaging user experience. (Gottardi et al. 2022)

  • Evaluate summarization metrics on multiple domains rather than just one, as performance may vary depending on the specific characteristics of the domain. (Sicong Huang, Celikyilmaz, and Li 2022)

  • Consider implementing a two-stage selector and reader for multi-hop question answering tasks, where the selector first selects the most relevant document to the question and then uses the question along with the selected document to find other relevant documents, followed by a reader that is initially trained on a single-hop QA dataset and then transferred to the multi-hop QA task. (X.-Y. Li, Lei, and Yang 2022)

  • Consider utilizing a prompt-based approach, specifically PromptDFD, for data-free knowledge distillation (DFKD) to enhance the quality of synthesized samples and achieve significant improvements in distillation performance. (Xinyin Ma et al. 2022)

  • Consider employing unsupervised sentence simplification techniques to improve the performance of machine reading comprehension (MRC)-based event extraction, particularly in cases involving long-range dependencies and syntactically complex sentences. (Mehta, Rangwala, and Ramakrishnan 2022)

  • Consider implementing a novel method of soft prompt tuning incorporating an additional soft prompt at decoder level of T5, which produced superior performance compared to the existing baseline model of T5 with a single encoder soft prompt, and demonstrated that the best classifier trained with artificial data produced from the proposed novel model, produces not just random classification results but interpretable results based on the different positive and negative words of the input text. (Senadeera and Ive 2022)

  • Develop a comprehensive understanding of the underlying mechanisms driving the phenomenon being studied before attempting to draw any conclusions or make recommendations. (Shaham et al. 2022)

  • Consider using large language models (LLMs) to generate situated robot task plans by leveraging programming language structures, enabling direct execution of generated plans, and incorporating situated state feedback from the environment. (I. Singh et al. 2022)

  • Utilize a comprehensive and rigorous assessment framework to accurately evaluate the planning and reasoning capabilities of large language models (LLMs), as opposed to relying solely on simplistic benchmarks that may not fully capture the complexity of these tasks. (Valmeekam et al. 2022)

  • Consider evaluating large language models like GPT-3 on a variety of analogy tasks, including text-based matrix reasoning, letter string analogies, four-term verbal analogies, and story analogies, to understand your emerging abilities in zero-shot reasoning. (Webb, Holyoak, and Lu 2022)

  • Consider the potential negative societal implications of large language models, particularly regarding your environmental impact and the lack of diverse perspectives in your development, and seek ways to mitigate these risks through collaborative, transparent, and responsible approaches. (Workshop et al. 2022)

  • Consider using large-scale contrastive language-audio pretraining models with feature fusion and keyword-to-caption augmentation to achieve superior performance in text-to-audio retrieval tasks and state-of-the-art performance in zero-shot audio classification tasks. (Yusong Wu et al. 2022)

  • Consider employing various methods to address the primary challenge faced by non-autoregressive translation (NAT) models - the failure to effectively capture target dependency - through data manipulation, modeling, training criterion, decoding, and leveraging pre-trained models. (Yisheng Xiao et al. 2022)

  • Consider using pre-trained language models (PLMs) for controllable text generation (CTG) tasks, as they offer improved interpretability and controllability compared to traditional deep learning methods, allowing for more accurate and reliable text generation. (Yunpeng Zhang et al. 2022)

  • Develop a unified model capable of simultaneously predicting an answer and providing an explanation, thereby improving efficiency and ensuring that the explanation remains closely tied to the reasoning process behind the answer. (H. Jeong et al. 2022)

  • Be aware of the evolving landscape of artificial intelligence (AI) tools using Large Language Models (LLMs) such as ChatGPT, and consider the potential impacts on academic integrity, particularly in terms of plagiarism and academic misconduct, when conducting studies in the field of education. (Abd-Elaal, Gamage, and Mills 2022)

  • Carefully consider the potential biases and limitations inherent in language models like ChatGPT, particularly regarding fairness, privacy, and accuracy, before implementing them in higher education settings. (Ya Dai et al. 2022)

  • Utilise a novel approach called subgraph reasoning’, which allows for improved explainability and increased accuracy in fake news detection by focusing on specific subgraphs within propagation networks.’ (Ruichao Yang et al. 2022)

  • Consider utilising large language models, such as GPT-3.x, to enhance the efficiency and accuracy of generating outage summaries for cloud systems. (Jia Chen, Wang, and Wang 2022)

  • Utilize the concept of knowledge neurons’ and apply a knowledge attribution method to identify these neurons in pretrained transformers, allowing for targeted editing of specific factual knowledge without requiring fine-tuning.’ (D. Dai et al. 2021)

  • Leverage heterogeneous data, study dependencies in societal events, and interpret event predictions to effectively utilize deep learning for societal event forecasting. (S. Deng and Ning 2021)

  • Adopt a multi-faceted approach to detecting and analyzing online hate speech, incorporating natural language processing, multimedia computer vision, and community detection techniques, while considering the broader socio-political context of extremism and radicalization. (L. Gao et al. 2021)

  • Consider utilizing customized pre-trained transformer models (Custom PTMs) pre-trained on app reviews when higher performance and lower prediction times are needed for app issue classification. (Hadi and Fard 2021)

  • Adopt a novel approach to evaluating large language models (LLMs) by treating them as participants in psychology experiments, rather than relying solely on traditional performance-based benchmarks, in order to gain deeper insight into your mechanisms of decision-making, reasoning, cognitive biases, and other crucial psychological aspects. (Hendrycks et al. 2021)

  • Consider multiple linguistic, quasi-linguistic, and training-related factors when investigating cross-linguality in shared embedding space, including word order agreement, morphological complexity agreement, and in-family training data. (Jones, Wang, and Mahowald 2021)

  • Consider both pipeline-based and joint-based event extraction paradigms when developing deep learning models for event extraction tasks, taking into account the potential for error propagation in pipeline-based methods and the benefits of simultaneous trigger and argument role classification in joint-based methods. (Qian Li et al. 2021)

  • Carefully consider the integration of Pseudo-Relevance Feedback (PRF) signals with deep language models, specifically focusing on balancing effectiveness and efficiency, and exploring various approaches such as text-based and vector-based PRF methods. (Hang Li et al. 2021)

  • Leverage existing pre-trained language models (PLMs) through fine-tuning for the task of interest, prompting the PLMs to perform the desired task, or reformulating the task as a text generation problem with application of PLMs to solve it accordingly, resulting in continuous establishment of new state-of-the-art performances. (B. Min et al. 2021)

  • Carefully construct prompts for large language models (LLMs) to effectively utilize your abilities in generating functional and secure code for vulnerability repair. (Pearce et al. 2021)

  • Focus on developing effective techniques for creating high quality text augmentations, carefully selecting appropriate negative samples, and leveraging the power of contrastive learning to enhance the performance of your natural language processing models. (Rethmeier and Augenstein 2021)

  • Carefully consider the choice of tuning method (prompt vs. prefix) when working with language models, as the results indicate that these methods exhibit varying degrees of domain robustness depending on factors like model size and prompt length. (Qinyuan Ye, Lin, and Ren 2021)

  • Carefully balance the scaling of model size and data size, aiming for a roughly 1:1 ratio to optimize performance for a given amount of training compute. (Barocas et al. 2021)

  • Consider implementing a “Topic Aware Sampling” (TAS) strategy when constructing training datasets for dense retrieval tasks. This involves clustering queries into groups based on shared thematic content, and then selecting representative samples from within these clusters to create diverse yet coherent training batches. This approach helps to maximise the information gained from each training iteration, leading to improved performance and reduced reliance on computationally expensive techniques such as large batch sizes or continuous index updates. (Hofstätter et al. 2021)

  • Employ a sentence-level topic analysis methodology to effectively track the spread of specific disinformation narratives across both news sites and social media platforms. (Angelov 2020)

  • Consider using a dual-tower transformer architecture with cross-attention for accurate synchronization of multiple channels in spoken dialogue modeling, leading to more naturalistic turn-taking and backchanneling. (Baevski et al. 2020)

  • Consider using a combination of rule-based and deep learning approaches for natural language search (NLS) in complex domains, such as Customer Relationship Management (CRM), to effectively balance accuracy, scalability, and explainability. (Borges et al. 2020)

  • Consider combining term-counting and machine learning methods to improve the accuracy of sentiment classification tasks, particularly when dealing with valence shifters like negations, intensifiers, and diminishes. (Daus, Ptashnyk, and Raithel 2020)

  • Consider incorporating knowledge-guided linguistic rewrites as a secondary source of evidence when generating inference rule corpora, as it can significantly improve the quality of the corpora and increase the accuracy of inferred facts. (P. Jain, Rathi, and Chakrabarti 2020b)

  • Consider incorporating a human-in-the-loop question corrector within your text-to-SQL systems to enhance robustness against untranslatable user inputs. (Kelkar et al. 2020)

  • Carefully examine and address potential topic bias in your models, particularly when working with smaller datasets, as doing so may lead to improved performance. (C.-S. Wu et al. 2020)

  • Develop general-purpose pretraining approaches tailored to learning representations for both natural language utterances and structured database tables, as demonstrated through the creation of TaBert, a pretrained language model that jointly learns representations for natural language sentences and (semi-)structured tables. (P. Yin et al. 2020)

  • Consider implementing a multi-stage method based on a hierarchical encoder-decoder model to explicitly model utterance-level attention distribution at training time, while enforcing diversity at inference time using a unigram diversity term. (Manakul, Gales, and Wang 2020)

  • Focus on developing unified embedding models that combine text, user, and context information to achieve improved recall in personalized search engines like Facebook. (J.-T. Huang et al. 2020)

  • Avoid using sampled metrics for item recommendation due to your inconsistencies with exact versions, and if sampling must be done, corrections can be applied to improve the quality of estimates. (Krichene and Rendle 2020)

  • Utilize both positive and negative evidential paths within a knowledge graph to accurately evaluate the truthfulness of a given factual statement. (Jiseong Kim and Choi 2020)

  • Consider utilising pre-training methods on large databases of relevant information to enhance the performance of recurrent neural networks in tasks such as URL segmentation. (Hao Zhang, Ro, and Sproat 2020)

  • Adopt a late interaction’ paradigm for information retrieval systems, which involves separate encoding of queries and documents into contextual embeddings, followed by efficient and pruning-friendly computations between both sets to determine relevance. (Khattab and Zaharia 2020)

  • Consider using TaPas, a weakly supervised question answering model that reasons over tables without generating logical forms, as it simplifies the architecture, enables pre-training, expands the range of question types handled, and supports direct handling of conversational settings. (Herzig et al. 2020)

  • Utilize the Electronic Database on Investment Treaties (EDIT) due to its comprehensive nature, uniformity, machine-readability, annotatability, and free accessibility, allowing for more accurate and thorough studies of international investment agreements. (Alschner, Elsig, and Polanco 2020)

  • Consider using text-free prompting methods instead of text-based ones when working with sign language translation tasks, as it leads to improved naturalness and comprehension in the resulting translations. (V. Kumar, Choudhary, and Cho 2020)

  • Carefully evaluate the trade-offs between model size, suitability for fast inference, and perplexity when selecting compression techniques for recurrent neural networks in language modeling, with matrix decomposition techniques often providing the best balance. (Grachev, Ignatov, and Savchenko 2019)

  • Develop a comprehensive understanding of the various factors influencing the relationship between variables before attempting to draw conclusions about causality. (Fetahu, Anand, and Koutraki 2019)

  • Utilise salience-awareness and discourse-profiling to effectively isolate the main event chains from other distracting events in natural language text, thereby improving narrative prediction and event-based temporal question answering. (Yinhan Liu et al. 2019)

  • Pay close attention to hyperparameter tuning and training data size when comparing different language model pretraining approaches, as these factors can significantly affect the final results. (Yinhan Liu et al. 2019)

  • Utilize data-dependent complexity (DDC) to assess the compatibility between text representations and tasks, providing a calibrated, quantitative measure of the difficulty of a classification-based NLP task, allowing for comparison between representations without needing empirical evaluations that might be influenced by initialization and hyperparameters. (Yinhan Liu et al. 2019)

  • Consider using a 3-part hinge loss function instead of a traditional 2-part hinge loss function when dealing with semantic matching problems in product search, as it allows for better distinction between random negative examples, impressed but not purchased examples, and positive examples (purchased items). (Nigam et al. 2019)

  • Aim to develop large-scale, multi-domain, grounded datasets that capture various aspects of conversational search, enabling the evaluation of all desired competencies of a conversational search system. (Penha, Balan, and Hauff 2019)

  • Consider using OpenHowNet, an open sememe-based lexical knowledge base built upon HowNet, for natural language processing tasks due to its comprehensive coverage of over 100,000 senses annotated with sememes, along with its user-friendly web interface and API for easy access and manipulation. (F. Qi et al. 2019)

  • Consider incorporating a two-stage process in your studies, wherein the first stage involves providing a CQA example alongside the corresponding CoS-E explanation to a language model, which is trained to generate the CoS-E explanation, and the second stage involves using the language model to generate explanations for each example in the training and validation sets of CQA, which are then provided to a second common-sense reasoning model by concatenating it to the end of the original question (Rajani et al. 2019)

  • Leverage the strong pattern matching behavior and dense representations of deep neural networks to create a mapping between test instances and training instances with known labels, allowing them to update the model by updating the data and labels in these mappings. (Schmaltz 2019)

  • Consider utilizing simple BERT-based models for relation extraction and semantic role labeling tasks, as they have demonstrated state-of-the-art performance without requiring external features like part-of-speech tags or dependency trees. (P. Shi and Lin 2019)

  • Utilize the MASS (Masked Sequence to Sequence pre-training) technique for language generation tasks, which involves pre-training the encoder and decoder jointly within an encoder-decoder framework, allowing for improved performance across various language generation tasks such as neural machine translation, text summarization, and conversational response generation. (K. Song et al. 2019)

  • Utilise a large-scale Transformer model consisting of three encoders: an object relationship encoder, a language encoder, and a cross-modality encoder, to effectively learn vision-and-language connections. (H. Tan and Bansal 2019)

  • Focus on developing small and efficient multilingual models for sequence labeling tasks, such as part-of-speech tagging and morphological prediction, through techniques like parameter sharing, cross-lingual transfer learning, and model distillation, resulting in improved accuracy, reduced system complexity, and enhanced applicability to multilingual or codemixed inputs. (Tsai et al. 2019)

  • Utilize the DocRED dataset when studying document-level relation extraction, as it is the largest human-annotated dataset for this purpose, requiring multi-sentence reading and synthesis of information across a document. (Yuan Yao et al. 2019)

  • Employ a two-stage decoding process in your natural language generation models, allowing the model to generate each word of the summary considering both sides context information, thereby enhancing the naturalness and coherency of the generated sequence.’ (Haoyu Zhang, Xu, and Wang 2019)

  • Employ an unsupervised end-to-end training scheme to discover discrete subword units from speech without using any labels, using an ASR-TTS autoencoder reconstruction setting, where an ASR-Encoder is trained to discover a set of common linguistic units given a variety of speakers, and a TTS-Decoder trained to project the discovered units back to the designated speech. (A. T. Liu, Hsu, and Lee 2019)

  • Focus on developing and implementing contextual embedding alignment procedures to enhance the performance of multilingual BERT models, particularly in terms of zero-shot transfer capabilities. (Aldarmaki and Diab 2019)

  • Focus on developing deep supervised generative models for predicting individual keyphrases, rather than relying solely on abstractive text summarization techniques for generating keyphrase strings. (Çano and Bojar 2019)

  • Develop a new loss function that has an inbuilt threshold to differentiate between random negative examples, impressed but not purchased examples, and positive examples (purchased items) when conducting semantic matching in product search. (Nigam et al. 2019)

  • Consider developing dedicated models and non-standard quality measures for the challenging task of abstractive dialogue summarization, as demonstrated through the creation and analysis of the SAMSum Corpus, a high-quality chatdialogues corpus manually annotated with abstractive summarizations. (Gliwa et al. 2019)

  • Consider employing a graph-based coarse-to-fine method when conducting unsupervised bilingual lexicon induction tasks, as it allows for better utilisation of clique-level information and reduction of noise in pre-trained embeddings. (Artetxe and Schwenk 2019)

  • Utilize the One Sense per Category (OneSeC) methodology for generating multilingual sense-annotated data, which involves using the association between Wikipedia pages and categories to map them onto word senses, resulting in improved performance in Word Sense Disambiguation (WSD) tasks. (Scarlini, Pasini, and Navigli 2019)

  • Focus on scaling up language models to improve your performance in few-shot learning tasks, as evidenced by the success of the 540-billion parameter Pathways Language Model (PaLM) in achieving state-of-the-art results on numerous language understanding and generation benchmarks. (Allamanis et al. 2018)

  • Consider utilising low rank matrix factorisation during training to compress the word embedding layer, which represents the size bottleneck for most NLP models. This method allows for recovery of accuracy while maintaining the reduced size, and outperforms alternative methods like fixed-point quantization or offline word embedding compression. (Acharya et al. 2018)

  • Focus on developing and refining analysis methods for neural networks in natural language processing, particularly those that address the challenges of interpretability, causality, and evaluation. (Belinkov and Glass 2018)

  • Consider leveraging pre-trained transformer-based language models, such as BERT, RoBERTa, and ALBERT, to address commonsense validation and explanation tasks, as they demonstrate promising performance across various subtasks. (Cer et al. 2018)

  • Utilize probing tasks to understand the linguistic properties captured within sentence embeddings, allowing for improved evaluation and comparison of various encoder architectures and training methods. (Conneau, Kruszewski, et al. 2018)

  • Consider adopting a “privacy by design” approach in the development of voice assistants, ensuring that user data remains securely on local devices rather than being transferred to external servers. (Coucke et al. 2018)

  • Utilise a Multiplex Graph Convolutional Network (Multi-GCN) to simultaneously model diverse types of relationships among sentences and words, thereby improving the efficiency and accuracy of extractive text summarisation. (Yue Dong et al. 2018)

  • Utilise causally sufficient embeddings - low-dimensional document representations that preserve enough information for causal identification and enable efficient estimation of causal effects - when dealing with high-dimensional textual data in order to make accurate causal inferences. (Egami et al. 2018)

  • Consider adopting a hierarchical annotation scheme for semantic parsing tasks, as it enables the representation of complex compositional queries, can be efficiently and accurately parsed by standard constituency parsing models, and outperforms sequence-to-sequence approaches on the released dataset. (Sonal Gupta et al. 2018)

  • Consider utilizing the WikiHow dataset for evaluating your summarization systems, as it offers a large-scale, diverse, and abstractive resource that differs significantly from traditional news-focused datasets. (Koupaee and Wang 2018)

  • Consider using a Search, Label, and Propagate’ (SLP) framework for bootstrapping intents from existing chat logs using weak supervision, which reduces hours to days of labeling effort down to minutes of work by leveraging a search engine to find examples and a data programming approach to automatically expand the labels. (Mallinar et al. 2018)

  • Utilise the Train-o-Matic tool to create large, high-quality sense-annotated corpora for various languages, enabling superior performance in Word Sense Disambiguation tasks, particularly in low-resource languages. (Pasini, Elia, and Navigli 2018)

  • Carefully choose your neural architecture (such as LSTM, CNN, or self-attention) when working with contextual word representations derived from pre-trained bidirectional language models (biLMs), as this decision affects both end-task accuracy and the nature of the learned representations. (Peters et al. 2018)

  • Consider using hyperbolic word embeddings instead of traditional Euclidean embeddings for improved performance in tasks like similarity, analogy, and hypernymy detection. (Tifrea, Bécigneul, and Ganea 2018)

  • Consider carefully selecting appropriate textual resources when training word embeddings for biomedical NLP applications, as different sources may lead to varying levels of accuracy and relevancy in capturing semantic properties and linguistic relationships between words. (Yanshan Wang et al. 2018)

  • Utilise a combination of semantic signals derived from both words and entities when encoding documents for effective representation learning. (Yamada, Shindo, and Takefuji 2018)

  • Consider using a combination of skip-gram models and hierarchical softmax techniques to effectively learn and represent complex feature signatures in large-scale heterogeneous information networks, allowing for improved clustering and identification of potentially fraudulent devices. (Chao Xu et al. 2018)

  • Focus on analyzing the linguistic properties, network architecture, and learning objectives of multilingual BERT (M-BERT) to fully understand its cross-lingual abilities. (Conneau, Rinott, et al. 2018)

  • Consider the impact of tokenization choices on the discovery of linguistic and evolutionary relationships between languages, as subword tokenization provides a strong bias towards knowledge of these relationships. (Artetxe, Labaka, and Agirre 2018)

  • Consider incorporating pseudo-relevance feedback techniques into your dense retrieval models to improve the accuracy and efficiency of your information retrieval processes. (J. Johnson, Douze, and Jégou 2017)

  • Consider utilizing distant supervision techniques combined with logistic regression and convolutional neural network classifiers for accurate identification of individuals killed by police within news article corpora. (Keith et al. 2017)

  • Consider combining multiple techniques when training distributed word representations, including position-dependent features, phrase representations, and subword information, to achieve higher quality results across various NLP tasks. (Mikolov et al. 2017)

  • Leverage refined alignment of latent representations to perform style transfer in non-parallel text corpora, assuming a shared latent content distribution across different text corpora. (T. Che et al. 2017)

  • Develop a fully data-driven, knowledge-grounded neural conversation model that generalizes the Sequence-to-Sequence (Seq2Seq) approach by conditioning responses on both conversation history and external “facts”. (Ghazvininejad et al. 2017)

  • Consider using a combination of gated attention-based recurrent networks and self-matching mechanisms to improve the effectiveness of your models in reading comprehension tasks. (Yichen Gong and Bowman 2017)

  • Consider using a hybrid learning objective that combines reinforcement learning with supervised learning to overcome exposure bias and improve the readability of generated summaries. (Paulus, Xiong, and Socher 2017)

  • Consider using a joint attribute-preserving embedding model when conducting cross-lingual entity alignment, as it effectively combines structure embedding and attribute embedding to create a unified vector space that improves the accuracy of identifying equivalent entities across different languages. (Zequn Sun, Hu, and Li 2017)

  • Consider utilising a deep LSTM encoder from an attentional sequence-to-sequence model trained for machine translation (MT) to contextualise word vectors, as this approach was demonstrated to significantly improve performance across a range of common NLP tasks including sentiment analysis, question classification, entailment, and question answering. (Hill et al. 2017)

  • Consider incorporating domain knowledge through keyword annotation when developing query-to-query reformulation solutions, as demonstrated by the QUEEN model, which significantly improved the performance of existing seq2seq and transformer models. (G. Klein et al. 2017)

  • Utilise the X-WikiRE dataset to explore the benefits of multilingual approaches in relation extraction, particularly focusing on cross-lingual transfer and simultaneous performance in multiple languages. (Cer et al. 2017)

  • Consider combining automated labelling techniques with human-labelled data to enhance the performance of event extraction systems. (Yubo Chen et al. 2017)

  • Utilize the UN General Debate corpus (UNGDC) as a valuable resource for deriving estimates of government preferences on various policy dimensions using text analytic methods, thereby enriching our understanding of international politics. (Baturo, Dasandi, and Mikhaylov 2017)

  • Utilize the newly introduced CySecED dataset for event detection in cybersecurity texts, as it offers a larger variety of event types, requires consideration of document-level context for accurate predictions, and provides a more realistic assessment of the challenges faced in this domain compared to previous datasets. (Lifu Huang et al. 2017)

  • Utilize the Self-Annotated Reddit Corpus (SARC) for sarcasm detection studies, as it offers a large, diverse, and self-annotated dataset that allows for accurate model training and evaluation across various scenarios. (Khodak, Saunshi, and Vodrahalli 2017)

  • Utilize deep contextualized word representations, like ELMo, which effectively model complex characteristics of word usage and variation across linguistic contexts, leading to significant improvements in various natural language processing tasks. (Xiaodong Liu et al. 2017)

  • Focus on developing a “Path Language Model” to identify salient and coherent event-event paths, which can then be merged into graph schemas to improve the accuracy of information extraction tasks. (Modi et al. 2017)

  • Carefully consider the choice of language model, including the size of the n-gram, smoothing technique, and evaluation metric, to effectively capture the nuances of human language and achieve optimal performance in tasks involving sequential data. (Neubig 2017)

  • Consider applying the Skip-gram with Negative Sampling (SGNS) technique, commonly referred to as word2vec, to item-based collaborative filtering problems, as it has demonstrated strong performance in natural language processing tasks and could potentially yield improved results in this domain. (Barkan and Koenigstein 2016)

  • Be mindful of the potential for gender bias in word embeddings, as demonstrated through the example of the w2vNEWS embedding, and implement debiasing techniques to mitigate this issue before utilizing these models in downstream applications. (Bolukbasi et al. 2016)

  • Leverage the power of learning-to-rank algorithms combined with personalization strategies, such as homophily and user intent analysis, to improve search results relevancy in complex, multi-vertical environments like LinkedIn. (Ha-Thuc and Sinha 2016)

  • Utilize a diverse and large-scale dataset like WikiReading to train and evaluate deep neural network models for natural language understanding tasks, as it provides a comprehensive benchmark for evaluating the effectiveness of various model architectures. (Hewlett et al. 2016)

  • Focus on developing large-scale language models using techniques such as character-level CNNs and softmax loss, and utilize large datasets like the One Billion Word Benchmark to improve the accuracy and efficiency of these models. (Jozefowicz et al. 2016)

  • Consider incorporating a termination state in your neural network architectures to allow for dynamic determination of reasoning depth, rather than relying on a fixed number of turns during inference. (Kadlec et al. 2016)

  • Consider utilizing machine learning techniques to analyze visual social media data, specifically focusing on color analysis, metadata components, and algorithmic face detection, in order to accurately identify markers of depression and potentially improve early screening and detection of mental illness. (Reece and Danforth 2016)

  • Consider integrating Large Language Models (LLMs) into your experimental processes, particularly within software startups, to potentially mitigate common inhibitors and improve overall efficiency and outcomes. (Unterkalmsteiner et al. 2016)

  • Utilize a supervised classification methodology with natural language processing (NLP) features to effectively detect abusive language in user comments, thereby improving upon existing deep learning approaches. (Nobata et al. 2016)

  • Utilize a comprehensive machine learning framework for ranking within categories, blend separate rankings in All Product Search, employ NLP techniques for matching queries and products, and develop algorithms specifically designed for unique tasks of certain categories. (Sorokina and Cantu-Paz 2016)

  • Consider using product quantization (PQ) for compressing text classification models, as it provides a balance between memory usage and accuracy, resulting in models that fit in limited memory spaces. (Joulin, Grave, Bojanowski, Douze, et al. 2016)

  • Consider utilizing the Stanford Question Answering Dataset (SQuAD) for developing and evaluating machine comprehension models, as it offers a large, high-quality, and diverse resource for testing the effectiveness of algorithms in handling complex linguistic structures and reasoning processes. (Rajpurkar et al. 2016)

  • Consider utilising an aspect-aware multimodal summarisation model when dealing with complex data sources such as those found in e-commerce product summarisation. This model allows for the integration of multiple types of data, such as visual and textual information, and determines the most salient aspects of a product, thus improving the overall effectiveness of the summarisation process. (Wenyuan Zeng et al. 2016)

  • Optimize your autoregressive flow-based generative network for text-to-speech synthesis by maximizing the likelihood of the training data, enabling a stable and effective system for controlling speech variation and style transfer. (Nishimura et al. 2016)

  • Leverage the power of learning-to-rank algorithms combined with personalization features extracted from user profiles and behaviors to achieve highly effective and tailored search experiences in professional networks. (Ha-Thuc and Sinha 2016)

  • Consider using word embeddings and vector space comparison to analyze smart contract code, allowing for effective identification of code clones, bug detection, and contract validation. (Bojanowski et al. 2016)

  • Use a two-stage approach to generate coherent long text, involving the creation of a directed semantic graph through clustering similar nodes of a family of document-level paths using a revised self-organizing map (SOM), followed by the extraction of maximum matching paths or subgraphs to ensure the preservation of extra but relevant content related to the short input text. (Kusner and Hernández-Lobato 2016)

  • Utilize a Recurrent Neural Network (RNN)-based sequence model named “SummaRuNNer” for extractive summarization of documents, as it demonstrates superior performance compared to existing state-of-the-art models and offers interpretability advantages due to its ability to visually break down predictions into abstract features such as information content, salience, and novelty. (Nallapati, Zhai, and Zhou 2016)

  • Consider using a Byte-to-Span (BTS) model, which leverages Long Short Term Memory (LSTM) models and sequence-to-sequence models to process text at the byte level, enabling accurate multilingual language processing without reliance on language-specific pipelines or external data sources. (Gillick et al. 2015)

  • Consider utilizing compositional vector space models for knowledge base completion, as they enable chains of reasoning on paths of any length, generalize to unseen paths, and facilitate zero-shot prediction of target relations without explicit training. (Neelakantan, Roth, and McCallum 2015)

  • Consider utilizing temporal convolutional neural networks (ConvNets) for text understanding tasks, as they can effectively learn hierarchical representations of words, phrases, and sentences from raw character sequences, potentially outperforming traditional approaches like bag-of-words or word2vec models. (Xiang Zhang and LeCun 2015)

  • Carefully examine and address potential gender biases in word embeddings before applying them in machine learning models, as these biases can perpetuate harmful stereotypes and negatively impact model performance. (Zliobaite 2015)

  • Employ a multi-perspective approach to sentence similarity modeling, leveraging convolutional neural networks to extract features at various granularity levels and using multiple types of pooling, ultimately leading to improved performance on paraphrase identification and semantic textual similarity tasks. (Hua He, Gimpel, and Lin 2015)

  • Focus on creating high-quality simplification datasets, like the Newsela corpus, to overcome the limitations of current simplification resources like Simple Wikipedia, thereby improving the reliability and validity of findings in text simplification studies. (Wei Xu, Callison-Burch, and Napoles 2015)

  • Consider utilizing the chargrid methodology for processing and understanding structured documents, as it preserves the 2D layout of these documents and enables accurate information extraction tasks. (Kai Chen et al. 2015)

  • Adopt a holistic approach to opinion spam detection by combining multiple sources of evidence, such as linguistic cues, behavioral footprints, and relational ties within a unified framework, allowing for improved accuracy and scalability. (Rayana and Akoglu 2015)

  • Utilise multi-agent coordination communication games to facilitate the development of artificial intelligence capable of engaging in interactive communication, specifically focusing on referential games where agents must develop a language to coordinate and earn payoffs. (Mikolov, Joulin, and Baroni 2015)

  • Use an adaptive metric approach like TransA for knowledge graph embedding tasks, as it provides a more flexible and effective way to handle complex entities and relationships compared to existing translation-based methods. (Han Xiao et al. 2015)

  • Optimize the bias-variance trade-off of the Pairwise Inner Product (PIP) loss to determine the optimal dimensionality for word embeddings, providing a theoretically sound solution to the open problem of dimensionality selection. (Bahdanau, Cho, and Bengio 2014)

  • Use distributed representations of sentences and documents, such as Word2Vec or Doc2Vec, which enable more accurate language understanding through contextual learning and improved performance on various natural language processing tasks like sentiment analysis and information retrieval. (Quoc V. Le and Mikolov 2014)

  • Focus on developing methods to predict hypernymy relationships in word embeddings, as opposed to simply classifying them, because this approach offers greater potential for understanding complex linguistic structures. (Vilnis and McCallum 2014)

  • Utilize a continuous cache model for improving neural language models, allowing for efficient adaptation to recent history and significant performance enhancement across multiple language model datasets. (Bahdanau, Cho, and Bengio 2014)

  • Utilize subgraph embeddings to enhance the accuracy of your models in answering complex questions from a knowledge base. (Bordes, Chopra, and Weston 2014)

  • Consider applying a structured event graph to guide sentence fusion in text summarization, allowing for the integration of both similar and disparate sentence fusion within a unified framework. (Bahdanau, Cho, and Bengio 2014)

  • Utilize topic- and author-controlled natural experiments on Twitter to isolate the impact of wording on message propagation while controlling for potential confounding variables such as author popularity and topic relevance. (Chenhao Tan, Lee, and Pang 2014)

  • Consider using gated convolutional networks for language modeling due to your superior performance, computational efficiency, and ability to handle long-term dependencies when compared to traditional recurrent neural networks. (Chelba et al. 2013)

  • Utilise the Paragraph Vector methodology for creating semantic representations of input sequences of varying lengths, as it outperforms traditional bag-of-words models and doesnt require specific tuning or reliance on parse trees.’ (Mikolov, Chen, et al. 2013)

  • Utilise standardised event-based representations and file formats when conducting information extraction tasks in the biomedical domain, enabling cross-task system reuse and facilitating comparisons between different approaches. (“UZH in BioNLP 2013” 2013)

  • Carefully consider the unique challenges posed by Twitter data, such as the presence of meaningless messages, spam, and informal language, and adapt your event detection techniques accordingly. (Atefeh and Khreich 2013)

  • Utilize Paragraph Vector, a novel method for generating semantic representations of variable-length input sequences, which outperforms traditional bag-of-words models and improves accuracy in various natural language processing tasks like sentiment analysis and text classification. (Mikolov, Chen, et al. 2013)

  • Consider leveraging the social network structure and extracting more than 20,000 features for each account to train supervised machine learning models that classify accounts across many different kinds of abuse. (R. Arora, Dekel, and Tewari 2012)

  • Focus on developing effective algorithms for learning to rank, taking into account factors like training and testing, data labeling, feature construction, evaluation, and relationships with ordinal classification. (LI 2011)

  • Consider using phase-based multilabel classification when analyzing large-scale datasets, particularly when faced with limited availability of labeled data, as it enables accurate and efficient analysis through the use of controlled vocabulary and crowdsourced labeling of frequently occurring phrases. (Bekkerman and Gavish 2011)

  • Carefully evaluate and compare various measures of graph connectivity when developing graph-based algorithms for unsupervised word sense disambiguation, as different measures may lead to varying levels of accuracy and performance. (Navigli and Lapata 2010)

  • Leverage both open-source and closed-source large language models (LLMs) to improve content-based recommendation systems, particularly by integrating open-source LLMs as content encoders and generating additional data via closed-source LLMs to enhance user profiles and content understanding. (Bu et al. 2010)

  • Incorporate both local and global context in your models to improve word representations and better capture the semantics of words, while still retaining syntactic information. (Collobert and Weston 2008)

  • Consider using weighted finite-state transducers as a common framework for representing and optimizing various models in speech recognition, as it offers significant algorithmic and software engineering benefits. (“Springer Handbook of Speech Processing” 2008)

  • Utilise a combination of syntax and semantics in your analysis of text to accurately extract political events, through the use of a rule-based system that leverages grammatical information and machine learning models trained on labeled spans to determine the correct event property for each span of words. (Blei and Lafferty 2007)

  • Consider utilizing multiple clause constructors in your Inductive Logic Programming (ILP) models to enhance the expressiveness and performance of your hypotheses. (“Proceedings of the 2nd French-Speaking Conference on Mobility and Uibquity Computing - UbiMob ’05” 2005)

  • Consider using Bayesian inference or reducing the number of states in Expectation Maximization (EM) algorithms when working with Hidden Markov Models (HMMs) for Part-of-Speech (POS) tagging tasks, due to your superior performance compared to traditional EM methods. (Alexander Clark 2003)

  • Consider applying instance-weighting at the level of phrase pairs, rather than at the sentence level, when working with statistical machine translation (SMT) systems. (Koehn, Och, and Marcu 2003)

  • Consider the heterogeneity of information across different sections of a scientific article when conducting large-scale extraction of particular items of information, as each section contains certain keywords that are unique to it. (Shah et al. 2003)

  • Utilise a show-and-tell procedure in which visual scenes are paired with natural language descriptions to train a spoken language generation system. (D. K. Roy 2002)

  • Distinguish between short-term and long-term repetition priming, as they involve different underlying mechanisms and have distinct implications for understanding the cognitive processes involved in word identification. (Bowers 2000)

  • Consider applying random walk Markov chain theory to measuring lexical semantic relatedness, as it allows for the principled combination of multiple types of edges from WordNet and the aggregation of local similarity statistics across the entire graph, leading to similarity scores that are highly correlated with human judgments. (A. Berger and Lafferty 1999)

  • Consider keeping exceptional training instances in memory for improved generalization accuracy in language learning tasks, as opposed to removing them through training set editing techniques or decision-tree learning methods. (Daelemans, Bosch, and Zavrel 1998)

  • Consider utilizing Amazon Mechanical Turk for collecting non-expert annotations due to its cost effectiveness and ability to produce high-quality results comparable to those obtained through traditional expert labeling methods. (C. F. Baker, Fillmore, and Lowe 1998)

  • Consider using a novel AST-based Neural Network (ASTNN) approach for source code representation, which involves splitting large ASTs into smaller statement trees and encoding them using a bidirectional RNN model to capture the naturalness of statements, ultimately producing a vector representation of a code fragment. (Hochreiter 1998)

  • Combine the use of machine-readable dictionaries and spreading and activation models to create very large neural networks (VLNNs) for effective word sense disambiguation (WSD) in natural language processing. (Veronis and Ide 1990)

  • Consider adopting the Plover ontology and Polecat dataset for improved accuracy and utility in analyzing political event data, due to its simpler structure, better integration of contextual information, and ease of updating through machine learning models. (Azar 1984)

  • Consider utilizing supervised discriminative machine learning techniques, specifically rank preference learning, to effectively model grade relationships between scripts in automated grading systems for ESOL exams. (Bloom 1970)

  • Carefully consider the statistical properties of your learning environments or datasets, as connectionist and other machine learning algorithms can be highly sensitive to these factors. (NA?)

  • Utilize a multi-layered Kohonen self-organizing feature map to effectively categorize internet homepages based on your content, thereby improving internet keyword searching and browsing. (NA?)

  • Focus on developing a trainable information extraction system that utilizes an ontology defining classes and relations of interest, alongside a set of training data containing labeled regions of hypertext representing instances of these classes and relations, in order to efficiently extract relevant information from webpages and hyperlinks. (NA?)

  • Consider using a combination of Webfoot and CRYSTAL systems to effectively extract meaningful information from diverse web page formats, achieving high levels of accuracy and reliability even in cases where traditional natural language processing techniques struggle due to the absence of clear syntax. (NA?)

  • Focus on developing a trainable information extraction system that utilizes an ontology and a set of training data to effectively extract information from web pages and hyperlinks, enabling the creation of a comprehensive and computer-understandable knowledge base. (NA?)

  • Focus on developing machine learning algorithms that can accurately predict user behavior on the web, specifically in terms of which hyperlinks users are likely to click on, in order to create more efficient and personalized online experiences. (NA?)

  • Consider conducting open challenge evaluations to objectively assess and compare the performance of different text mining systems in assisting biological database curation processes, thereby demonstrating measurable progress in the field. (NA?)

  • Focus on developing natural language dialogue (NLD) facilities for tutoring environments involving verbal and qualitative subject matters, where the shared knowledge between the tutor and the learner is low to moderate, and use expectation- and misconception-tailored (EMT) dialogue strategies to effectively advance the dialogue and learning agenda. (NA?)

  • Consider implementing an automated extraction algorithm, such as MuteXt, to efficiently identify and extract mutation data from the vast amounts of scientific literature, thereby facilitating the integration of this information into existing databases and improving overall understanding of protein families. (NA?)

  • Incorporate simple semantics into Topic Detection and Tracking (TDT) by splitting the term space into groups of terms that share the same type of meaning, such as locations, proper names, temporal expressions, and general terms. These groups can then be associated with an external ontology to determine the similarity of two terms within the same group, allowing for better organization of news documents according to news events. (NA?)

  • Utilise an integrated system like GeneWays to efficiently and accurately extract and manage information on molecular interactions from large volumes of journal articles. (NA?)

  • Consider using discriminative models for information retrieval tasks due to your ability to handle complexities such as modeling assumptions, expressiveness, learning arbitrary features, and the explicit notion of relevance. (NA?)

  • Utilize label propagation, specifically Modified Adsorption, to effectively classify the polarity of tweets by leveraging the relationships between tweets, authors, and features in a graph-based framework. (NA?)

  • Focus on achieving a balance between high precision and reasonable recall when developing text mining applications for extracting protein annotations from biomedical literature, taking into account the complexity of GO terms and protein names, and utilizing diverse strategies such as pattern matching, machine learning, and template extraction. (NA?)

  • Consider utilizing the Structural Semantic Interconnection (SSI) algorithm for Word Sense Disambiguation (WSD) tasks, as it effectively combines multiple lexical resources and employs a context-free grammar to identify relevant semantic patterns between them, leading to accurate sense classification. (NA?)

  • Utilize the Dirichlet Compound Multinomial (DCM) model rather than the conventional multinomial model for text analysis tasks, due to the formers ability to accurately capture the ‘burstiness’ phenomenon inherent in natural language.’ (NA?)

  • Utilize Explicit Semantic Analysis (ESA) to improve the correlation of computed relatedness scores with human judgments in the realm of semantic relatedness of natural language texts. (NA?)

  • Utilize a combination of knowledge-based methodologies and statistical methods to effectively mine biomedical literature for new, potentially causal connections between biomedical terms. (NA?)

  • Utilize reinforcement learning techniques within a Markov Decision Process (MDP) framework to optimize dialogue strategies in spoken dialogue systems, while considering the trade-offs between model-based and simulation-based approaches to strategy learning. (NA?)

  • Consider utilizing test collections as a valuable resource for evaluating retrieval systems, enabling rapid, reproducible experiments in a controlled environment without requiring users. (NA?)

  • Focus on developing efficient methods for extracting relevant information from web pages, such as utilizing an initial portion of a web page for analysis, to improve the performance of your machine learning models. (NA?)

  • Utilize text-mining and information extraction strategies to efficiently access and analyze the vast amounts of biological information contained in online scientific literature collections, thereby streamlining the entire research process from experiment planning to result interpretation and communication. (NA?)

  • Focus on developing automatic ontology construction techniques that integrate and extend existing knowledge sources, such as Wikipedia and WordNet, to improve various applications including advanced query processing, faceted browsing, automated infobox edits, and template generation. (NA?)

  • Utilize a combination of manual labelling and machine learning approaches to effectively detect and mitigate the impact of opinion spam in online reviews. (NA?)

  • Consider multiple measures of semantic richness - including number of semantic neighbours, number of features, and contextual dispersion - when investigating the impact of semantic richness on visual word recognition, as each measure contributes uniquely to response time and error variance in lexical decision and semantic categorisation tasks. (NA?)

  • Utilize a computational modelling framework that defines intermediate semantic features based on co-occurrence statistics of input stimulus words within a large text corpus, and trains the model via multiple regression against observed fMRI images to derive maximum likelihood estimates for model parameters. (NA?)

  • Adapt boosting techniques for information retrieval tasks by incorporating LambdaRank gradients in the training process, allowing them to optimize non-smooth IR metrics like NDCG and improve efficiency without sacrificing accuracy. (NA?)

  • Utilise clustering techniques to discover exemplar terms that ensure comprehensive semantic coverage of a document, prior to extracting key phrases from said document. (NA?)

  • Utilize a combination of multi-feature spaces, active learning strategies, and ensemble methods to improve the efficiency and accuracy of biomedical text classification tasks, specifically in the context of citation screening for systematic reviews. (NA?)

  • Consider utilizing a multi-component approach to accurately classify Twitter users into human, bot, and cyborg categories, incorporating entropy-based, machine-learning-based, account property-based, and decision-making components. (NA?)

  • Utilize a probabilistic graphical model to simultaneously make decisions regarding entity, type, and relation annotations for web tables, resulting in increased accuracy and efficiency. (NA?)

  • Consider using probabilistic latent semantic analysis (PLSA) to identify the main topics in a corpus of documents, followed by selecting words and phrases that are strongly associated with only one or a few topics as the terms to include in a term map, thus reducing subjectivity and labor intensity compared to traditional manual methods. (NA?)

  • Utilize Sentic Computing, a multidisciplinary approach combining computer and social sciences, to overcome limitations of traditional opinion mining and sentiment analysis methods by recognizing and processing implicit content within texts. (NA?)

  • Consider incorporating social network information into your sentiment analysis models, as it can lead to statistically significant improvements in user-level sentiment classification accuracy. (NA?)

  • Consider using smartphones as a more natural and convenient alternative to traditional body sensors for activity recognition tasks, especially considering the growing prevalence of smartphone usage among younger populations. (NA?)

  • Adapt a machine learning-based system for the identification and extraction of potential adverse drug event relations from MEDLINE case reports, utilizing a high-quality corpus that was manually annotated using an ontology-driven methodology. (NA?)

  • Consider implementing semantic interaction in your visual analytic tools, which involves enabling analysts to interact directly within the visual metaphor using actions derived from your analytic process, such as searching, highlighting, annotating, and repositioning documents. This approach can lead to improved accuracy and efficiency in the analysis of complex datasets. (NA?)

  • Consider both the content and authorship of online chatter when studying its effects on consumer behavior, as different types of messages and authors can have varying levels of persuasion and influence. (NA?)

  • Consider incorporating feature similarity into your Vector Space Models (VSMs) through the use of soft cosine measures and soft similarity, rather than solely relying on traditional hard cosine measures and independence assumptions. (NA?)

  • Utilize a multi-faceted approach to knowledge base construction, combining noisy extractions from diverse web sources with prior knowledge derived from existing knowledge repositories, and leveraging supervised machine learning methods for fusing these disparate information sources. (NA?)

  • Adopt a multi-faceted approach to transform word embeddings to the sense level and leverage knowledge from a large semantic network for effective semantic similarity measurement, achieving state-of-the-art performance on multiple datasets. (NA?)

  • Focus on developing machine learning models that utilize low-cost features for accurate identification of fake Twitter followers, rather than relying solely on high-performing but expensive features. (NA?)

  • Exercise caution when using pre-trained language models to study human language processing, as larger transformer-based models tend to memorize sequences during training, causing your surprisal estimates to diverge from humanlike expectations. (NA?)

  • Utilize literature mining to efficiently extract specific types of experimental evidence for drug-drug interactions (DDIs), focusing on pharmacokinetic evidence, which is crucial for understanding causal mechanisms and guiding future pharmacological and epidemiological investigations. (NA?)

  • Aim to develop an independent relation extraction system that minimizes reliance on supervised NLP modules for features, thereby reducing error propagation and improving overall performance. (NA?)

  • Use a combination of weak classifiers to accurately predict information from curated data, followed by a cost-sensitive learner to improve the accuracy of the prediction. (NA?)

  • Explore the application of deep learning techniques to improve the performance of dialogue systems, particularly in areas such as language understanding, dialogue state tracking, policy learning, and natural language generation. (NA?)

  • Combine machine learning techniques with cognitive behavioral modeling to effectively identify and distinguish between human and bot participants in social media disinformation campaigns, allowing for deeper understanding of your respective roles and interactions. (NA?)

  • Adopt a perturbation-driven paradigm when analyzing and interpreting natural language inference models, allowing them to dynamically explore the interactions between different components of the model and better understand its mechanisms. (“Distill,” n.d.)

  • Develop specialized sentiment analysis tools tailored to the unique characteristics of developer communication channels, such as Stack Overflow, to avoid misclassifications caused by generic sentiment analysis tools trained on non-technical domains. (NA?)

  • Consider implementing a unified LLM-based dialogue management module in your conversational recommender systems, which simplifies the architecture and enhances flexibility, while addressing the challenges of control and guidance towards a reasonable dialogue policy. (NA?)

  • Explore the efficacy of different prompt engineering methods for knowledge extraction, utilizing a relation extraction dataset in conjunction with a large language model (such as GPT-4), and propose a novel evaluation framework grounded in Wikidata ontology to address the challenge of evaluation. (NA?)

  • Use prompt engineering strategies, such as chain-of-thought prompting and prompt chaining, to effectively leverage large language models like ChatGPT for accurate and efficient data analysis and interpretation. (NA?)

  • Employ deep learning techniques like recurrent neural networks and transfer learning, particularly using sent2affect, to achieve significant improvements in text-based emotion recognition for decision support. (NA?)

  • Utilise advanced statistical methods and machine learning techniques to analyse large datasets of social media interactions in order to understand the role of social bots in the spread of misinformation. (NA?)

  • Consider the role of social influence and the characteristics of the messenger when analyzing the impact of online word-of-mouth communication on product sales. (NA?)

  • Consider utilizing unsupervised word embeddings to extract latent knowledge from large volumes of scientific literature, as demonstrated effectively in the field of materials science. (NA?)

  • Consider employing cross-modal representation learning to enable seamless querying across diverse data formats within multi-modal data lakes, thereby reducing the need for extensive data integration efforts. (NA?)

  • Focus on developing efficient methods for augmenting language models with massive-scale memory, rather than solely increasing model size and training data volume. (NA?)

  • Consider employing multiple exploratory, descriptive, and classification techniques in conjunction with each other to maximize the value derived from textual data, particularly in the context of rapidly developing fields like AI and NLP. (NA?)

  • Utilise the Neural Tangent Kernel (NTK) approach to better understand the dynamics of fine-tuning pre-trained language models, particularly in low-data settings. (NA?)

  • Adopt a constructive design science approach to explore the use of large language models (LLMs) in software maintenance, focusing on creating a framework for prompt engineering that offers systematic guidance for improving software maintenance processes. (NA?)

  • Consider using a novel story generation method which involves dividing the story creation process into three distinct stages: inputting the start of the narrative, generating potential events through a common sense reasoning model, filling these events into a question template to create queries, employing a question answering model to produce responses, selecting the response with the lowest perplexity score as part of the story, and repeating this cycle until a full story is produced. This method has been shown to be capable of producing more coher (NA?)

  • Consider using flexible prompt templates and self-learning strategies to improve the accuracy, transfer-ability, and generality of your models for software requirement classification and auto-labeling. (NA?)

  • Carefully choose among different extraction pattern methods depending on the characteristics of your application domains, considering factors such as syntactic/semantic constraints, delimiter-based approaches, and the ability to generate multi-slot rules. (NA?)

  • Use an iterative prompt refinement technique to improve the performance of ChatGPT in predicting gene relationships, which involves assessing prompt efficacy using metrics like F-1 score, precision, and recall, then engaging GPT-4 to suggest improved prompts. (NA?)

  • Consider using a direct prompting strategy instead of more complex ones like the chain of thoughts (CoT) or modified CoT methods when working with ChatGPT (GPT-3.5) for medical problem-solving tasks, as it demonstrates non-inferior accuracy and simplifies interaction with the model. (NA?)

  • Consider employing prompt engineering strategies like chain-of-thought and problem separation to enhance the performance of large language models like ChatGPT in pharmacokinetic data analysis tasks, despite persistent issues with numerical accuracy and reproducibility. (NA?)

  • Adopt a three-layered approach to auditing large language models, combining governance audits, model audits, and application audits in a structured and coordinated manner to effectively manage ethical and social risks. (NA?)

  • Employ the GPEI methodology, consisting of four steps: define the objective, design the prompt, evaluate the response, and iterate, to optimize interactions with AI-language models like ChatGPT in the field of engineering. (NA?)

  • Focus on developing long-answer prompt learning methods (such as KLAPrompt) to effectively integrate structured semantic knowledge into pre-trained language models, leading to significant improvements in understanding Chinese word semantics and medical science. (NA?)

  • Adopt a narrative and critical literature review approach to examine the various aspects of AI prompt engineering as a digital competence, considering its development, technological challenges, and applications across different domains. (NA?)

  • Carefully consider the potential benefits and risks of using ChatGPT in healthcare education, practice, and research, taking into account factors such as ethical concerns, data privacy, and the possibility of incorrect or inaccurate information. (NA?)

  • Exercise caution while using ChatGPT for literature synthesis, citations, problem statements, research gaps, and data analysis in academic research, despite its effectiveness in initial idea generation. (NA?)

  • Utilise two complementary approaches to assess the impact of training data volume on model-to-human alignment: evaluating GPT-2 models trained on varying dataset sizes against an fMRI benchmark, and testing the performance of a GPT-2 model trained on a 9-billion-token dataset at various stages throughout its training process. (NA?)

  • Consider employing domain-specific vocabulary and pretraining when working with large neural language models for biomedical natural language processing tasks, as these approaches can enhance model robustness and improve performance. (NA?)

  • Carefully consider the unique characteristics of educational data, such as its hierarchical structure and discrete nature, when selecting appropriate data augmentation techniques like large language models for generating synthetic datasets that effectively represent varying levels of prior knowledge and students difficulties.’ (NA?)

  • Consider utilizing large language models like ChatGPT for medical education assistance, as they demonstrate strong performance on the United States Medical Licensing Exam (USMLE) and offer valuable insights through your explanations. (NA?)

  • Employ a combination of generative and associative approaches when developing creativity support tools for journalists, allowing them to efficiently explore various angles and perspectives while maintaining trust and credibility in your reporting. (NA?)

  • Utilise the k-Class Directed Preferential Attachment (kCDPA) model when attempting to identify fake accounts in social media platforms, particularly during the early stages of account creation. (NA?)

  • Focus on developing multimodal pre-training methods that leverage large-scale unsupervised image-text pairs, specifically targeting entity and relational information within text and image, to effectively align the semantic spaces of images and text, ultimately improving the extraction of multimodal entities and relations. (NA?)

  • Consider developing a unified generative retriever (UGR) for knowledge-intensive language tasks (KILTs), which can efficiently handle various retrieval tasks at different levels of granularity through n-gram-based identifiers and prompt learning strategies. (NA?)

  • Focus on developing question-specific prompts for large language models to accurately generate structured answers to legal questions, while maintaining effectiveness when generalized prompt templates are used instead. (NA?)

  • Actively engage with large language models (LLMs) such as ChatGPT, leveraging your intelligence, versatility, and collaboration capabilities to increase research productivity, while remaining mindful of ethical concerns and promoting transparency in your use. (NA?)

  • Carefully consider the limitations of using large language models like ChatGPT when evaluating your performance in specific domains, such as medical exams, due to potential biases and lack of up-to-date information. (NA?)

  • Utilise a mixed methods approach, combining social network analysis, content analysis of interviews, and investigation of user experiences, to gain a holistic understanding of the concerns surrounding the use of chatbots, particularly ChatGPT, in education. (NA?)

  • Develop a comprehensive taxonomy of neural text-to-SQL systems to facilitate comparisons, identify challenges, and guide future research in the field. (NA?)

  • Carefully consider the differences between human communication and artificial intelligence systems, particularly large language models (LLMs), to accurately assess your capabilities and limitations, and avoid attributing human-like qualities to them unnecessarily. (NA?)

Computer Vision

  • Consider fine-tuning a pre-trained model instead of training a model from scratch to erase undesirable concepts from text-to-image diffusion models, using the models own knowledge and no additional data.’ (Gandikota et al. 2023)

  • Consider implementing the Instance-Centroid Faster Point Sampling Module (IC-FPS) in your point-based 3D object detection models to improve efficiency and accuracy in large-scale point cloud scenes. (Haotian et al. 2023)

  • Consider using a Bayesian probabilistic resolution to prompt learning for vision-language models, where label-specific stochastic prompts are generated hierarchically by first sampling a latent vector from an underlying distribution and then employing a lightweight generative model. Additionally, the authors suggest semantically regularizing prompt learning with visual knowledge and viewing images and corresponding prompts as patch and token sets under optimal transport, pushing the prompt tokens to faithfully capture the label-specific visual concepts, instead (Xinyang Liu et al. 2023)

  • Adopt an iterative action research approach to explore the solution space of text-to-image generation, focusing on subject terms and style modifiers, and utilizing a four-stage process of initial prompt, composition adjustment, style refinement, and variation selection to create believable illustrations of pre-existing texts. (Ruskov 2023)

  • Combine flow-based SR model with the local implicit module to effectively address the ill-posed nature of SR and solve the arbitrary-scale challenge. (J.-E. Yao et al. 2023)

  • Consider implementing unsupervised prompt learning (UPL) for vision-language models to avoid laborious prompt engineering and improve transfer performance without relying on labeled data. (Yihong Huang et al. 2022)

  • Preserve the pre-trained weights of existing diffusion models and add new trainable gated Transformer layers to enable the injection of new grounding information, thereby allowing for greater controllability and improved quality in text-to-image generation. (Sheynin et al. 2022)

  • Consider implementing a bi-directional coding-based end-to-end stereo image compression network (BCSIC-Net) to effectively reduce inter-view redundancy and outperform state-of-the-art methods in stereo image compression tasks. (H. Ma et al. 2022)

  • Utilise the HANA database to test the robustness of your handwritten text recognition (HTR) methods and models on more challenging, large-scale, and highly unbalanced databases. (C. M. Dahl et al. 2021)

  • Separate “skills” and “concepts” in Visual Question Answering (VQA) models to improve generalization to out-of-distribution data, and proposes a novel method for doing so through implicit learning of grounded concept representations and disentangled encoding of skills and concepts. (Hendricks et al. 2021)

  • Consider incorporating the CTC-Prefix-Score during S2S decoding in order to improve the accuracy of your handwritten text recognition models. (Wick, Zöllner, and Grüning 2021)

  • Consider combining patch-level and line-level approaches in historical document classification tasks, particularly for script and font classification and document dating, as this combination significantly improved results in the ICDAR 2021 Competition on Historical Document Classification. (“Document Analysis and Recognition – ICDAR 2021” 2021)

  • Incorporate various types of contextual information, including both interlocutors, scene, and task information, when attempting to infer personality in dyadic scenarios, as doing so leads to significant improvements in accuracy. (Palmero et al. 2020)

  • Extend the stopping method based on next integrated recognition result modelling to be used within a string result recognition model with per-character alternatives, as it achieves higher accuracy compared to previous methods based on input observations clustering. (K. Bulatov, Savelyev, and Arlazarov 2020)

  • Consider using Gated Fully Convolutional Networks (GFCNs) for unconstrained handwritten text recognition tasks, as they offer a recurrence-free alternative to traditional CNN+LSTM architectures, resulting in faster training and prediction times with fewer parameters, while maintaining competitive accuracy levels. (Moysset and Messina 2019)

  • Consider using the Structured Skip List (SSL) data management method for real-time indoor 3D reconstruction tasks, as it combines the benefits of both ordered and unordered methods while improving storage efficiency and operation efficiency. (S.-J. Li et al. 2018)

  • Focus on improving labels in supervised learning systems, specifically making them soft, informative, collective, and dynamic, to enhance overall system performance. (Bagherinezhad et al. 2018)

  • Utilize semantic network interpretation alongside traditional visualization techniques to gain a comprehensive understanding of a networks decision-making processes and improve overall model performance.’ (P. Guo and Farrell 2018)

  • Carefully select appropriate sensor modalities and apply advanced machine learning techniques such as Random Forest and AdaBoost to achieve accurate classification of affective states in individuals. (P. Schmidt et al. 2018)

  • Consider utilising domain adaptation techniques, particularly those involving deep convolutional architectures, to improve the accuracy of your models when dealing with differences in data distributions across various domains. (Csurka 2017)

  • Ensure that your datasets are balanced and representative of the full range of phenotypic variations in order to avoid biased results and unfair outcomes in automated facial analysis algorithms. (Dehghan et al. 2017)

  • Carefully consider the type of data required for your study, including the selection of appropriate sheet music samples and the creation of comprehensive ground truth annotations, in order to effectively evaluate and improve optical music recognition (OMR) systems. (Hajič and Pecina 2017)

  • Consider employing 2D Self-organized Operational Neural Networks (Self-ONNs) in conjunction with deformable convolutions to achieve improved accuracy in Handwritten Text Recognition tasks, as evidenced by the reduction in Character Error Rate (CER) and Word Error Rate (WER) observed in the study. (Puigcerver 2017)

  • Consider using the Maximum Mean Discrepancy (MMD) distance between local distributions of small patches in two images as a simple yet effective metric for comparing images, rather than relying solely on complex deep neural networks. (Arjovsky, Chintala, and Bottou 2017)

  • Utilize Conditional Generative Adversarial Networks (CGAN) to enhance the naturalness and diversity of image descriptions, thereby improving the overall quality of the generated sentences. (B. Dai et al. 2017)

  • Leverage A-la-carte Prompt Tuning (APT) to efficiently address the a-la-carte learning problem, enabling accurate and flexible model construction based on user-selected data sources. (Priya Goyal et al. 2017)

  • Use PointMixup, an interpolation method that generates new examples through an optimal assignment of the path function between two point clouds, to effectively apply data augmentation techniques to point cloud data. (Ravanbakhsh, Schneider, and Poczos 2016)

  • Adopt a class prototypes based supervised contrastive learning approach when dealing with fine-grained multilabel classification problems, particularly in the context of educational videos. (Abu-El-Haija et al. 2016)

  • Utilise a multi-scale, shared-net Fully Convolutional Neural Network (FCN) for improved text block detection, followed by a cascaded instance segmentation approach for accurate separation of word instances in arbitrary orientations. (Tong He et al. 2016)

  • Focus on improving the efficiency of deep convolutional neural networks by reducing redundancy and increasing information flow through techniques like Knowledge Distillation, adding higher-dimensional hint layers, incorporating output variances, and leveraging hand-crafted features. (J. Shen et al. 2016)

  • Utilize object-level grounding to establish a semantic link between textual descriptions and image regions, allowing for a more accurate and nuanced understanding of visual question answering tasks. (Karol Gregor et al. 2015)

  • Strive to understand the capabilities and limitations of specific machine-learning algorithms, considering factors like sample complexity, computational complexity, and the impact of modeling assumptions, in order to effectively leverage these tools for accurate and efficient learning from large datasets. (Horvitz and Mulligan 2015)

  • Consider incorporating gaze information in your studies, as it can significantly improve the accuracy and efficiency of video summarization tasks, particularly in cases involving egocentric videos. (Yeung, Fathi, and Fei-Fei 2014)

  • Consider using the solution path algorithm to optimize loss functions with ({0}) norm constraints, as it can help avoid the difficulties associated with directly minimizing the ({0}) norm or relying on rough approximations provided by the (_{1}) norm. (C. Lu and Tang 2014)

  • Utilise a combination of deep convolutional neural networks (CNNs) and recurrent neural networks (RNNs) to develop a single joint model capable of accurately translating images into coherent, descriptive sentences. (Vinyals, Toshev, et al. 2014)

  • Consider the impact of imbalanced data on performance metrics, particularly for facial action unit detection, and recommend reporting skew-normalized scores alongside the obtained ones to mitigate potential biases. (Jeni, Cohn, and Torre 2013)

  • Utilise a combination of computational materials discovery tools, unsupervised machine-learning algorithms, and density functional theory calculations to effectively navigate and understand the vast and complex landscape of inorganic ternary metal nitrides. (Müllner 2011)

  • Carefully consider the choice of statistical methods and model comparisons when analyzing complex datasets such as those derived from supernova observations, taking into account factors such as the dimensionality of the parameter space, the nature of the underlying physical processes, and the potential impact of nuisance parameters. (Elgarøy and Multamäki 2006)

  • Focus on developing a comprehensive Bayesian framework for using relevance feedback to guide a search, incorporating an entropy-minimizing display algorithm to maximize information gained from the user at each iteration, utilizing hidden annotations to improve accuracy and consistency, and employing experimental paradigms to quantitatively evaluate the performance of the system. (I. J. Cox et al. 2000)

  • Consider employing multiple attribute coding techniques, rule creation strategies, and bidding systems in combination to optimize the performance of your adaptive classifier systems. (NA?)

  • Utilize Fourier Transform Infrared Spectroscopy combined with machine learning techniques like Partial Least Squares Regression, Genetic Algorithms, and Genetic Programming to quickly and accurately detect microbial spoilage in beef. (NA?)

  • Use a combination of different imaging strategies, automated object detection, and feature subset selection to create a robust and fully automated framework for phenotype classification that achieves high accuracy even for applications with large biological variability and a considerable number of artifacts. (NA?)

  • Consider employing a neurofuzzy system for facial expression analysis, allowing for further learning and adaptation to specific users facial expression characteristics, thereby enhancing the robustness of the system across various individuals.’ (NA?)

  • Consider using Artificial Neural Networks (ANNs) like ARTMAP for large area land-cover modification mapping due to its high accuracy, resistance to training data deficiencies, and ability to perform equally well across diverse study areas with minimal human intervention. (NA?)

  • Take advantage of temporal coherence in unlabelled video data to enhance the performance of deep learning algorithms in object recognition tasks. (NA?)

  • Employ the similarity transformation to convert complex partial differential equations into simpler ordinary differential equations, followed by solving them numerically using the fourth-order Runge-Kutta integration scheme along with the shooting method. (NA?)

  • Consider employing machine learning techniques like model tree ensembles (MTE) to upscale FLUXNET observations of carbon dioxide, water, and energy fluxes to the global scale, thereby improving the prediction of site-level gross primary productivity (GPP), terrestrial ecosystem respiration (TER), net ecosystem exchange (NEE), latent energy (LE), and sensible heat (H) based on remote sensing indices, climate and (NA?)

  • Consider combining Object-based image analysis (OBIA) with Support Vector Machines (SVM) for enhanced vegetation separability, specifically for distinguishing between true and associated mangrove species, leading to improved classification accuracy. (NA?)

  • Carefully consider the choice of semiotic modality when instructing human subjects to perform gestures for data collection, as it significantly impacts the performance of the resulting gesture recognition system in terms of correctness and coverage. (NA?)

  • Consider utilizing deep convolutional neural networks (DCNNs) for quark/gluon jet discrimination, as they provide a promising alternative to traditional physically-motivated observables, potentially offering superior performance without relying heavily on prior knowledge or assumptions. (NA?)

  • Conduct a comprehensive comparison of major machine learning algorithms using the same land cover and land use classification scheme and the same satellite image to ensure accurate and reliable results. (NA?)

  • Consider utilising multi-modal approaches when analysing behavioural expressions, particularly combining facial actions and vocal prosody, to enhance the accuracy of depression detection. (NA?)

  • Consider utilizing deep convolutional neural networks for hyperspectral image classification due to your ability to achieve better classification performance than traditional methods like support vector machines and conventional deep learning-based approaches. (NA?)

  • Leverage deep learning algorithms, specifically deep convolutional neural networks like AlexNet and GoogLeNet, to accurately classify plant diseases using large, publicly available image datasets, potentially revolutionising smartphone-based crop disease diagnostics globally. (NA?)

  • Consider utilizing high-density surface electromyography (HD-sEMG) imaging techniques for analyzing muscle activity patterns, as they offer superior accuracy and efficiency compared to traditional methods. (NA?)

  • Utilize a data-driven modeling approach, combining inverse modeling and machine learning, to develop accurate and reliable turbulence models for predicting turbulent flow over airfoils. (NA?)

  • Utilise deep Convolutional Neural Networks (CNNs) for the classification of multispectral remote sensing images, particularly for complex land cover mapping like wetlands, as these networks offer superior classification accuracies compared to conventional machine learning tools like Random Forest and Support Vector Machines. (NA?)

  • Use a combination of architectural environments and virtual reality technology to effectively manipulate and measure emotional responses in participants. (NA?)

  • Consider employing various machine learning models like Random Forest, Extreme Gradient Boosting, and Deep Learning to effectively predict PM2.5 concentrations using multisource remote sensing data, while carefully evaluating feature importance to enhance model performance. (NA?)

  • Consider utilizing a combination of structural and spectral information derived from UAV remote sensing technology in conjunction with machine learning techniques to enhance the accuracy of your biomass estimation models. (NA?)

  • Consider utilising a semi-structured elicitation procedure, such as the game of charades, to collect more naturalistic gesture data in a controlled environment, enabling the investigation of individual differences in modes of representation and the development of effective gesture recognition algorithms. (NA?)

  • Utilize data-driven, quantitative techniques to overcome the challenge of defining the scope of digital humanities, allowing for a more accurate and comprehensive analysis of the field. (NA?)

  • Avoid relying on traditional random cross-validation methods when working with spatially correlated data, as it can lead to overconfidence in model predictions and misleading conclusions. (NA?)

  • Consider the appropriateness of the form of decision support based on the nature of the task, as some forms might require more cognitive effort and time investment, while others might be simpler and more effective. (NA?)

  • Leverage deep neural networks, specifically discriminative and generative networks, to efficiently address the forward and inverse problems in photonics, thereby enabling rapid prototyping and optimization of photonic devices. (NA?)

  • Pay particular attention to the number of fonts used in creating synthetic datasets, as well as employing correction LSTM to reduce errors in predictions, in order to significantly boost the accuracy of non-Latin scene text recognition. (NA?)

  • Focus on developing semi-supervised learning strategies for remote sensing data analysis, utilizing both labeled and unlabeled data to improve the generalization capabilities of models. (NA?)

  • Focus on developing interactive lifelog retrieval systems that enable fast and effective access to multimodal lifelogs through benchmarking exercises like the Lifelog Search Challenge (LSC). (NA?)

  • Consider implementing a multi-objective evolutionary algorithm to improve the fidelity of generative AI outputs to user preferences, by utilizing a pre-trained generative model as an implicit mutation operation to create Pareto-optimized images. (NA?)

  • Adopt a dual-methodological approach combining autoethnography and online ethnography to deeply understand the nuances of prompt engineering and text-to-image art generation. (NA?)

  • Employ a data-driven approach to understand user behaviour patterns and preferences in order to design more intuitive and effective interfaces for text-to-image prompt engineering. (NA?)

Speech Recognition

  • Use the NoRefER metric to analyze the attention mechanisms of QE metrics in order to improve the explainability of ASR systems. (Javadi et al. 2024)

  • Consider using a sequence-to-structure generation paradigm combined with a conditioned generation method that leverages speech recognition transcripts as contextual cues when attempting to extract semantic events from speech signals. (J. Kang et al. 2024)

  • Focus on developing a robust data mining pipeline to create a large-scale training dataset of diverse music audio clips, each paired with multiple descriptive text labels, to ensure the production of high-quality samples from a deep generative model. (Qingqing Huang et al. 2023)

  • Consider utilising transformer models in speech processing tasks, as they offer improved performance in comparison to traditional recurrent neural networks, particularly when dealing with long-range dependencies and large datasets. (S. Latif et al. 2023)

  • Consider using prompt-conditioning fine-tuning to enhance your speech recognition models domain sensitivity, leading to significant reductions in word error rate across various domains.’ (F.-T. Liao et al. 2023)

  • Consider utilizing a neural codec language model (such as VALL-E) for text-to-speech synthesis, as it offers significant improvements in speech naturalness and speaker similarity when compared to conventional methods. (Chengyi Wang et al. 2023)

  • Consider leveraging large amounts of unpaired multilingual speech and text data alongside smaller amounts of transcribed data to train a single large universal ASR model, potentially leading to improved performance across various languages. (Yu Zhang et al. 2023)

  • Consider using automatically-generated transcriptions from publicly-available pre-trained ASR models to increase the size of your training sets, leading to significant reductions in Word Error Rate (WER) in audio-visual speech recognition tasks. (P. Ma et al. 2023)

  • Leverage joint speech-text representation learning to develop massively multilingual, zero supervised speech, automated speech recognition (ASR) models, which can significantly enhance ASR performance for languages with limited or no manually transcribed speech. (Zhehuai Chen et al. 2022)

  • Consider using adversarial speaker-consistency learning (ASCL) to improve the quality and speaker similarity of zero-shot multi-speaker text-to-speech (ZSM-TTS) systems. (B. J. Choi et al. 2022)

  • Leverage large-scale pretraining datasets like Multilingual LibriSpeech and VoxPopuli, and employ pre-training and fine-tuning methodologies to achieve state-of-the-art performances in speech recognition, speech language identification, and speech-text retrieval tasks. (Conneau et al. 2022)

  • Employ a two-stage approach combining early biasing and decoder biasing techniques to improve the accuracy of CTC models in recognising rare and out-of-vocabulary words. (Dingliwal et al. 2022)

  • Consider implementing a modular hybrid autoregressive transducer (MHAT) model for improved text-only adaptation in end-to-end speech recognition systems, as it allows for efficient adaptation of the internal language model (ILM) to text-only data without negatively impacting other model components. (Z. Meng et al. 2022)

  • Consider using self-supervised pre-training methods for building spoken language understanding (SLU) systems, particularly in low-resource scenarios, as they tend to produce stronger semantic and acoustic representations than supervised learning methods. (Y. Peng et al. 2022)

  • Consider using a joint training approach with both speech-text paired inputs and text-only unpaired inputs, rather than pre-training and fine-tuning, to improve the performance of your end-to-end (E2E) models in automatic speech recognition (ASR) tasks. (Sainath et al. 2022)

  • Carefully examine the trade-off between the quantity and quality of labeled data required to achieve optimal performance in end-to-end (E2E) models for automated speech recognition (ASR) applications, particularly in underrepresented domains like Air Traffic Control (ATC). (Zuluaga-Gomez et al. 2022)

  • Consider leveraging machine learning techniques to automate the coding of political campaign video advertisements, thereby improving efficiency and potentially achieving comparable levels of accuracy to traditional manual coding approaches. (Tarr, Hwang, and Imai 2022)

  • Consider using a non-autoregressive convolutional neural model for speech synthesis with explicit pitch and duration prediction, as demonstrated by TalkNet, which offers improved performance, reduced computational complexity, and faster inference speeds compared to autoregressive models. (Beliaev and Ginsburg 2021)

  • Consider adopting a non-autoregressive neural text-to-speech model with a fully differentiable duration model, which allows for automatic learning of token-frame alignments and token durations without requiring supervised duration signals, leading to improved efficiency and naturalness in synthetic speech production. (Elias et al. 2021)

  • Consider generating 48 kHz waveforms instead of 16 kHz or 24 kHz ones, as it provides higher perception quality and naturalness in speech synthesis. (Yanqing Liu et al. 2021)

  • Consider implementing a novel multi-encoder learning (MEL) method when working with transformer-based end-to-end speech recognition, as it allows for increased robustness and generalization without adding extra complexity during inference. (Lohrenz, Li, and Fingscheidt 2021)

  • Consider using a combination of phone synchronous decoding (PSD) algorithm, blank label deweighting approach, deep feedforward sequential memory network (DFSMN) layers, CNN-based stateless predictor, and singular value decomposition (SVD) technology to develop a highly-efficient speech recognition model on edge devices. (Yuekai Zhang, Sun, and Ma 2021)

  • Consider using the Heuristic Error Assignment Training (HEAT) approach instead of the widely used Permutation Invariant Training (PIT) approach for training end-to-end multi-talker speech recognition models, as HEAT is more computationally efficient and achieves higher accuracy. (Liang Lu et al. 2021)

  • Leverage a non-streaming ASR model as a teacher to generate transcripts on an arbitrarily large data set, which is then used to distill knowledge into streaming ASR models, resulting in significant reductions in word error rates (WER) for RNN-T models on multiple datasets and languages. (Doutre et al. 2020)

  • Consider using insertion-based models for end-to-end automatic speech recognition tasks, as they offer advantages such as generating an arbitrary generation order of an output sequence and being able to perform non-autoregressive output token generation without requiring additional components or heuristics to estimate the output token sequence length. (Fujita et al. 2020)

  • Consider utilising a combination of multi-condition training, semi-supervised learning, and transcription strategies to maximise the efficiency and accuracy of your wake word spotting models, particularly when dealing with limited data resources. (Yixin Gao et al. 2020)

  • Consider implementing “asynchronous revision” as an inference technique to unify streaming and non-streaming speech recognition models, allowing for dynamic latency adjustments and improved accuracy. (Mingkun Huang et al. 2020)

  • Consider implementing a combination of shallow and cold fusion methods for integrating external neural network language models (NNLMs) into recurrent neural network transducer (RNN-T) models, leading to significant improvements in word error rate (WER) reductions while maintaining the systems streamability, flexibility, and lightweight properties.’ (Suyoun Kim et al. 2020)

  • Consider comparing multiple end-to-end (E2E) models for large-scale speech recognition tasks, including RNN-Transducer (RNN-T), attention-based encoder-decoder (AED), and Transformer-AED, across both non-streaming and streaming modes, using substantial amounts of training data to ensure robustness and validity. (Bohan Li et al. 2020)

  • Consider integrating the state reuse chunk-SAE and the MTA based SAD into your online CTC/attention architecture for improved efficiency and performance in online speech recognition. (Haoran Miao et al. 2020)

  • Carefully examine the alignment behavior of the attention function in high-performance sequence-to-sequence models, and consider adding an additional constraint loss to improve the models capability for streaming inference.’ (T.-S. Nguyen et al. 2020)

  • Utilise a sequence-level emission regularisation approach called “FastEmit” in order to improve the speed and accuracy of streaming Automatic Speech Recognition (ASR) systems. This method works by applying latency regularisation directly onto the per-sequence probability in training transducer models, thereby enabling faster and more accurate predictions without requiring additional word alignment information from an existing model. (Jiahui Yu, Chiu, et al. 2020)

  • Consider implementing a unified framework called Dual-mode ASR, which allows for the simultaneous training of both streaming and full-context Automatic Speech Recognition (ASR) models. This approach leads to improved latency and accuracy in streaming ASR, particularly when combined with weight sharing and joint training of full-context ASR, along with in-place knowledge distillation during the training process. (Jiahui Yu, Han, et al. 2020)

  • Utilise the AlignTTS approach when developing efficient feed-forward text-to-speech systems without explicit alignment. This involves using a Feed-Forward Transformer to generate mel-spectrum from a sequence of characters, determining the duration of each character via a duration predictor, and applying an alignment loss during training to consider all possible alignments using dynamic programming. This method was found to deliver superior performance compared to Transformer TTS, and significantly increased efficiency. (Zhen Zeng et al. 2020)

  • Leverage open speech corpora in multiple languages to develop a few-shot transfer learning method for keyword spotting, enabling accurate identification of keywords with minimal training data. (Bluche and Gisselbrecht 2020)

  • Use adversarial learning techniques to separate speaker characteristics from linguistic representations in non-parallel voice conversion tasks, thereby achieving higher similarity in the converted voices. (J.-X. Zhang, Ling, and Dai 2020)

  • Utilize a multi-task self-supervised approach to learn speech representations, which involves using a single neural encoder followed by multiple workers solving various self-supervised tasks, thereby enabling the discovery of general, robust, and transferable features. (McFee et al. 2019)

  • Focus on developing self-supervised pre-training methods for speech recognition, specifically utilizing vq-wav2vec for quantization and BERT for representation learning, as this approach leads to significant improvements in performance even when working with minimal amounts of labeled data. (Baevski, Auli, and Mohamed 2019)

  • Focus on combining self-supervised context prediction tasks with discrete unit discovery methods to effectively learn discrete representations of speech, enabling the direct application of natural language processing algorithms to speech data. (Baevski, Schneider, and Auli 2019)

  • Consider using multi-speaker ClariNet, a fully end-to-end speech synthesis model, to generate high-quality speech from multiple speakers, as it outperforms state-of-the-art systems in terms of naturalness due to its ability to jointly optimize the entire model. (J. Park et al. 2019)

  • Consider using a feed-forward network based on Transformer for generating mel-spectrogram in parallel for Text to Speech applications, as it addresses the challenges of slow inference speed, lack of robustness, and limited controllability faced by autoregressive models. (Yi Ren et al. 2019)

  • Consider combining the alignment capabilities of the connectionist temporal classification (CTC) approach with the modelling strength of the attention mechanism in order to improve end-to-end automatic speech recognition (ASR) performance. (Moritz, Hori, and Roux 2019)

  • Integrate parameter pruning (PP) and parameter quantization (PQ) techniques to achieve a more compact deep learning-based speech enhancement (SE) model, balancing denoising performance and computational cost. (J.-Y. Wu et al. 2019)

  • Consider implementing a combination of disentanglement mechanisms in your self-supervised learning frameworks, specifically focusing on disentanglement in teachers, disentanglement in students, and speaker conditioning, to effectively disentangle speaker variations without significant content loss. (Kameoka et al. 2019)

  • Utilize a novel reference encoder architecture in your prosody transfer systems to capture temporal prosodic representations that are robust to source speaker leakage, thereby enhancing the overall performance of the system. (Lorenzo-Trueba et al. 2019)

  • Consider applying SpecAugment, a simple yet effective data augmentation technique, to enhance the performance of automatic speech recognition systems. (D. S. Park et al. 2019)

  • Consider using a two-stage inference method called “one-in-a-hundred” (OAH) in your hybrid CTC and attention models for streaming speech recognition tasks, as it allows for efficient generation of multiple candidate sequences followed by selection of the best candidate based on acoustic encoded states, resulting in a potential 20% reduction in character error rate compared to baseline CTC models. (Z. Tian et al. 2019)

  • Consider utilising end-to-end (E2E) models for on-device speech recognition, particularly those based on recurrent neural network transducers (RNN-T), as they offer significant advantages in terms of latency and accuracy over conventional CTC-based models. (Yanzhang He et al. 2018)

  • Consider implementing sequence-level knowledge distillation for model compression of attention-based sequence-to-sequence speech recognition, as demonstrated through achieving up to 9.8x parameter reduction with minimal accuracy loss of up to 7.0% word-error rate increase. (Mun’im, Inoue, and Shinoda 2018)

  • Consider using a “subtractive” definition of prosody, which involves extracting prosody from ground truth speech audio while accounting for variations due to phonetics, speaker identity, and channel effects. (S. Arik et al. 2017)

  • Consider implementing a two-stage CTC process in your speech recognition systems, whereby a preliminary system generates a noisy letter sequence, which is subsequently refined by a secondary system trained to consume this noisy sequence and produce a cleaner version. (Zweig et al. 2017)

  • Focus on optimizing the expected word error rate (WER) during acoustic model training for speech recognition, as it leads to significant improvements in WER over traditional methods like state-level minimum Bayes risk (sMBR) training. (Shannon 2017)

  • Utilize deep neural networks for each component of your text-to-speech system, simplifying the overall process and increasing flexibility compared to traditional methods that require extensive feature engineering and domain expertise. (S. Arik et al. 2017)

  • Consider utilising direct acoustics-to-word Connectionist Temporal Classification (CTC) models for automatic speech recognition (ASR) tasks, as they offer potential improvements in efficiency and accuracy over traditional methods, although they may require larger amounts of training data. (Audhkhasi et al. 2017)

  • Consider utilizing online sequence-to-sequence models for noisy speech recognition tasks, as these models can provide accurate real-time transcriptions while being relatively simple to implement and adaptable to varying levels of noise. (C.-C. Chiu, Lawson, et al. 2017)

  • Consider combining structural and optimization improvements to attention-based encoder-decoder architectures like Listen, Attend, and Spell (LAS) for significant performance enhancements in speech recognition tasks. (C.-C. Chiu, Sainath, et al. 2017)

  • Consider combining multiple techniques, such as CTC, attention mechanisms, and RNN-LM, to enhance the performance of end-to-end automatic speech recognition systems. (Hori et al. 2017)

  • Consider combining the strengths of different models through multitask learning, such as integrating CTC and SCRF models for speech recognition tasks, to achieve improved overall performance. (Liang Lu et al. 2017)

  • Consider extending end-to-end speech recognition frameworks to include multichannel speech enhancement techniques, specifically focusing on optimizing the entire inference process, including beamforming, based on the final ASR objectives like WER/CER. (Ochiai et al. 2017)

  • Consider implementing a network of deep neural networks for distant speech recognition, where all components are jointly trained and cooperate with each other through full communication, to overcome limitations in current systems such as lack of matching and communication between speech enhancement and speech recognition modules. (Ravanelli et al. 2017)

  • Consider utilizing very deep convolutional networks in order to enhance the expressive power and generalization capabilities of end-to-end automatic speech recognition (ASR) models, thereby potentially leading to improved accuracy and reduced word error rates. (Yu Zhang, Chan, and Jaitly 2016)

  • Consider implementing label smoothing and coverage promotion techniques to improve the performance of sequence-to-sequence models in speech recognition tasks. (Chorowski and Jaitly 2016)

  • Utilise a joint CTC-attention model within a multitask learning framework for end-to-end speech recognition to enhance robustness, achieve rapid convergence, and mitigate alignment issues. (Suyoun Kim, Hori, and Watanabe 2016)

  • Adopt a meta-learning approach for adaptive text-to-speech (TTS) with few data, utilizing a multi-speaker model with a shared conditional WaveNet core and independent learned embeddings for each speaker, enabling rapid adaptation to new speakers with minimal data. (Yonghui Wu et al. 2016)

  • Consider implementing a “quantization aware” training process that applies a proposed quantization scheme during network training, allowing them to recover most of the loss in accuracy introduced by quantization. (Alsharif et al. 2015)

  • Utilise an Attention-based Recurrent Sequence Generator (ARSG) in your end-to-end Large Vocabulary Continuous Speech Recognition System (LVCSR) to improve accuracy and efficiency. (Bahdanau, Chorowski, et al. 2015)

  • Consider utilising convolutional neural networks (CNNs) in speech recognition tasks, as they offer superior performance compared to traditional deep neural networks (DNNs) due to your unique structure that allows for greater invariance to small shifts in speech features along the frequency axis. (Abdel-Hamid et al. 2014)

  • Focus on developing end-to-end deep learning architectures for speech recognition, leveraging large amounts of diverse data and advanced training techniques, such as multi-GPU computation and data synthesis, to achieve superior performance compared to traditional pipeline approaches. (A. Hannun et al. 2014)

  • Focus on developing and evaluating bi-directional recurrent deep neural networks (BRDNNs) for first-pass large vocabulary continuous speech recognition tasks, as they offer significant improvements in character error rate (CER) and word error rate (WER) compared to traditional hidden Markov model (HMM) based systems. (A. Y. Hannun et al. 2014)

  • Consider utilizing a layer-by-layer learning strategy for training a multi-layer generative model of patches of speech spectrograms, which can lead to efficient compression of speech and potentially enhance scalable speech recognition or rapid speech content retrieval. (L. Deng et al. 2010)

  • Consider utilizing a combination of short-term features like Spectral Flatness (SF) and Short-term Energy for developing a robust and effective Voice Activity Detection (VAD) algorithm. (Homayounpour and Moattar 2009)

  • Carefully manipulate spectral and temporal cues in speech recognition tasks to determine your relative importance in understanding speech, particularly in challenging listening environments. (Q.-J. Fu, Chinchilla, and Galvin 2004)

  • Explore diverse techniques including automatic speech recognition, keyword spotting, sub-word indexing, and speaker identification to improve audio information retrieval and make audio less opaque. (NA?)

  • Focus on developing robust, non-standard methods for measuring dysphonia in Parkinsons disease patients, while combining them with traditional harmonics-to-noise ratio measures, to improve the accuracy of telemonitoring applications.’ (NA?)

  • Consider using the Target Approximation (TA) model when studying tone and intonation in Mandarin and English, as it effectively links articulatory mechanisms to higher-level processes in speech, providing a comprehensive understanding of how pitch targets are implemented across syllables. (NA?)

  • Carefully select appropriate emotional speech corpora, considering factors such as language, number of emotions, and collection method, to ensure accurate and reliable results in speech emotion recognition. (NA?)

  • Focus on understanding the differences between various machine learning paradigms, such as generative vs. discriminative learning, supervised vs. unsupervised learning, and adaptive vs. multi-task learning, in order to effectively apply them to the specific challenges of automatic speech recognition. (NA?)

  • Develop a common evaluation framework, including tasks and databases, to assess and compare the state-of-the-art algorithms in order to determine the significance of your progress and guide future research directions. (NA?)

  • Use high-density electrocorticography (ECoG) recordings to detect when participants hear or say an utterance and then decode its identity, while dynamically updating the prior probabilities of each answer using the decoded question likelihoods as context, leading to significant improvements in answer decoding. (NA?)

  • Consider utilizing a neural codec language model (such as VALL-E) for text-to-speech synthesis, as it offers significant improvements in speech naturalness and speaker similarity when compared to conventional methods. (NA?)

Robotics

  • Transform Neural Radiance Fields (NeRFs) into Poisson Point Processes (PPPs) to allow for rigorous quantification of uncertainty in NeRFs, particularly for computing collision probabilities for a robot navigating through a NeRF environment. (Timothy Chen, Culbertson, and Schwager 2023)

  • Combine the strengths of large language models (LLMs) for long-horizon and semantic reasoning with grounded models for local and embodiment grounding in order to effectively plan and execute long-horizon robotic tasks. (Wenlong Huang et al. 2023)

  • Consider incorporating game design principles into your AI development processes, as demonstrated by the SimBot Challenge, which combines elements of game design with AI development to create an engaging and dynamic environment for users to interact with simulations through screen-enabled devices. (H. Shi et al. 2023)

  • Leverage few-shot prompts with large language models to autoregressively generate low-level control commands for robots without task-specific fine-tuning, thereby enabling effective interaction with the physical environment. (Y.-J. Wang et al. 2023)

  • Consider using Dense Voxel Fusion (DVF) as a sequential fusion method for generating multi-scale dense voxel feature representations, which can improve expressiveness in low point density regions and enhance multi-modal learning by training directly with projected ground truth 3D bounding box labels, ultimately leading to improved 3D vehicle detection performance. (Mahmoud, Hu, and Waslander 2022)

  • Leverage recent advancements in neural radiance fields (NeRFs) to provide single-step visual foresight for a control barrier function (CBF)-based controller, enabling the filtering out of unsafe actions and the preservation of safety in vision-based control systems. (M. Tong, Dawson, and Fan 2022)

  • Consider using Neural Grasp Distance Fields (NGDF) for grasp learning, as it enables interpreting the implicit function as a cost, resulting in a more efficient and effective optimization process for grasp poses. (T. Weng et al. 2022)

  • Consider leveraging valuable information from past data, specifically data collected in past traversals of the same scene, to provide rich contextual information for disambiguating challenging cases in 3D object detection. (Y. You et al. 2022)

  • Consider leveraging the behaviors of other agents to create more diverse driving scenarios without needing to collect additional data, while addressing the challenges associated with partial observability through the use of supervisory tasks to learn an intermediate representation that is invariant to the viewpoint of the controlling vehicle. (Filos et al. 2021)

  • Consider utilising a novel problem formulation for perceiving, acting, and specifying goals with Transformers, specifically within the context of robotic manipulation. (Jaegle et al. 2021)

  • Consider utilizing self-aligning implicit representations of local surfaces to enable effective affordance transfer across various object categories. (Zhenyu Jiang et al. 2021)

  • Aim to develop end-to-end trainable models for joint detection and tracking tasks, rather than relying on traditional tracking-by-detection pipelines that involve heuristic matching steps. (Chenxu Luo, Yang, and Yuille 2021)

  • Consider using Isaac Gym, a high-performance GPU-based physics simulation platform, to improve the efficiency and effectiveness of your robot learning projects. (Makoviychuk et al. 2021)

  • Consider integrating RGB sensors into Lidar-based 3D recognition systems to improve the visibility of distant or small objects, thereby enhancing safety in autonomous driving applications. (T. Yin, Zhou, and Krähenbühl 2021)

  • Utilise machine vision, neural networks, and a 1W power laser to effectively neutralise mosquitoes, thereby reducing the risk of disease transmission. (Ildar 2021)

  • Consider using lane segments as proposals for intention modeling in autonomous driving systems, as they offer several advantages such as explicit intention representation, better interaction capture, and sharing capabilities among agents. (Y. Chai et al. 2019)

  • Consider using a hierarchical planning approach for complex tasks like Vision-Language Navigation (VLN), which involves breaking down the problem into smaller, manageable subtasks and then solving those subtasks sequentially. (Nachum et al. 2019)

  • Adopt a structured workflow when integrating machine learning into networking, involving steps such as problem formulation, data collection, data analysis, model construction, model validation, deployment, and inference. (Mowei Wang et al. 2018)

  • Utilise a novel neural network architecture called a Phase-Functioned Neural Network (PFNN) for real-time character control mechanisms. This network computes weights through a cyclic function that uses the phase as an input, alongside user controls, the previous state of the character, and the geometry of the scene. It is trained end-to-end on a large dataset of locomotion activities like walking, running, jumping, and climbing, and can automatically produce high (Holden, Komura, and Saito 2017)

  • Consider using a 3D convolutional neural network (CNN) for shape completion in robotic grasp planning, as it enables rapid runtime shape completion and can improve grasping performance. (Varley et al. 2016)

  • Use the Unscented Kalman Filter (UKF) on Lie Groups for Visual Inertial Odometry because it offers a combination of accuracy, robustness, and versatility while maintaining computational efficiency. (Barrau and Bonnabel 2015)

  • Utilise Lagrangian relaxation to transform complex, hard constraints into soft constraints, thereby simplifying the overall problem and enabling easier resolution. (Butt and Collins 2013)

  • Consider developing a low-cost, lightweight, and energy-efficient gesture-based interaction device called GesturePod, which can be attached to any white cane to help individuals with visual impairments perform smartphone tasks through gestures, thus improving task completion times and reducing reliance on touch-based interactions. (Ashbrook, Baudisch, and White 2011)

  • Utilize machine learning algorithms alongside surface electromyography (EMG) technology to improve the control and functionality of advanced hand prosthetics. (Castellini and Smagt 2008)

  • Consider utilizing Neural Time Fields (NTFields) for robot motion planning in cluttered environments, as it offers a physics-informed neural model driven by the Eikonal equation, allowing for efficient and accurate pathfinding without the need for extensive expert motion trajectory data. (Bohlin and Kavraki 2000)

  • Utilise a support vector machine-based methodology to accurately identify and distinguish various finger movements through surface electromyography (EMG) signals. (NA?)

  • Consider using artificial evolution to effectively synthesize controllers for complex systems like the swarm-bot, as it allows for the discovery of simple but effective controllers that exhibit scalability and generality across various group sizes and configurations. (NA?)

  • Treat autonomous navigation as a software problem, focusing on developing intelligent driving software rather than designing exotic vehicles. (NA?)

  • Incorporate social psychology perspectives when studying the acceptance and adoption of domestic robots, considering factors such as social interactions, institutions, and hierarchies, as well as subjective consumer perceptions of what robots are, how they work, and what they are capable of doing in a domestic environment. (NA?)

  • Carefully choose appropriate policy gradient methods for motor primitive learning in robotics, considering factors like scalability, parameterization, and safety, and avoiding statistical bias or generation of infeasible policies. (NA?)

  • Focus on minimizing the amount of prior knowledge required to formulate fitness functions in evolutionary robotics, thereby enabling the development of novel control strategies. (NA?)

  • Consider developing a comprehensive model for assessing the acceptance of assistive social agents among elderly populations, taking into account both functional and social dimensions of acceptance. (NA?)

  • Carefully consider the type of model being used (forward, inverse, mixed, or multi-step prediction) and choose appropriate learning control architectures (direct modeling, indirect modeling, or distal teacher learning) depending on the specific requirements of the robot control scenario. (NA?)

  • Consider employing Bayesian exploration techniques to optimize the selection of exploratory movements during the identification of textures, thereby improving overall classification performance. (NA?)

  • Consider utilizing a mixture of motor primitives (MoMP) algorithm to effectively combine multiple movement primitives in order to efficiently adapt to various situations and enhance the performance of robotic systems. (NA?)

  • Focus on developing and evaluating novel robotic exoskeletons for gait rehabilitation in stroke survivors, emphasizing safety, usability, and effectiveness in promoting patient engagement and recovery. (NA?)

  • Consider integrating autonomous research systems (ARES) into your workflows to improve efficiency and accuracy in materials research, particularly in cases involving complex, high-dimensional parameter spaces. (NA?)

  • Carefully consider the active perception paradigm when developing intelligent agents, as it enables the integration of perception and action in achieving goals, and involves selecting what to sense, how to sense it, when and where to do so, and why it is necessary. (NA?)

  • Consider employing the Real2Sim2Real pipeline for efficient and effective training of grasping networks, leveraging state-of-the-art neural surface reconstruction methods to generate high-quality meshes from real-world point clouds, thereby enabling faster and more accurate simulations. (NA?)

  • Integrate a nominal system with an additive nonlinear part of the dynamics modeled as a Gaussian Process (GP) in your model predictive control (MPC) approach, enabling cautious control through the direct assessment of residual model uncertainty. (NA?)

  • Develop a series of incrementally challenging embodied Turing tests based on various species, using rich behavioural datasets and detailed biochemical measurements to inform the design of AI systems that can control virtual animals to replicate your in vivo counterparts behaviours. (NA?)

Autonomous Vehicles

  • Consider incorporating Large Language Models (LLMs) into the decision-making processes of autonomous vehicles to enhance your flexibility, responsiveness, and overall performance in complex, real-world scenarios. (Can Cui et al. 2023)

  • Use dynamic probabilistic networks (DPNs) to represent and update the belief state of an autonomous vehicle, enabling it to make informed decisions in real-time, even in the face of noisy and partial observational data. (Knott et al. 2023)

  • Consider incorporating a shared control architecture in your designs, which involves a collaboration between a human pilot and an adaptive autopilot, to enhance the resilience of flight control systems in the face of anomalies. (Sklar and Sarter 1999)

  • Explore the potential of reinforcement learning (RL) for developing autonomous controllers for cooperative adaptive cruise control (CACC) systems, as it offers a scalable and flexible approach that can effectively manage high-dimensional systems without requiring explicit knowledge of the underlying Markov decision process (MDP). (NA?)

  • Carefully select and combine various driving style recognition algorithms, such as rule-based, model-based, and machine learning methods, while taking into consideration the specific application requirements and constraints, to achieve accurate and effective driving style identification for improved vehicle energy management, driving safety, and advanced driver assistance systems. (NA?)

  • Use a cyber-physical system (CPS)-based co-design optimization approach to develop adaptive control algorithms for automated electric vehicles, taking into account various driving styles and your effects on vehicle dynamics, drivability, and energy efficiency. (NA?)

  • Consider the potential for adversarial attacks on LiDAR-based perception systems in autonomous vehicles, particularly through strategic control of spoofed points to fool machine learning models, and develop appropriate defenses accordingly. (NA?)

Game Playing Agents

  • Follow a systematic and iterative Design Science Research (DSR) approach, using the adapted process model proposed by Kuechler and Vaishnavi (2008), to develop a comprehensive understanding of virtual companionship and create a robust design theory for it. (Strohmann et al. 2022)

  • Leverage deep learning and natural language processing techniques to create a novel evaluation function for Chess moves, which is pre-trained on sentiment of commentary associated with the training moves, and guides and optimizes the agents game-playing decision making. (Kamlish, Chocron, and McCarthy 2019)

  • Focus on developing a novel training method centered around the concept of performing relative comparisons between positions, allowing for the creation of a significantly larger training dataset and ultimately leading to improved chess program performance. (David, Netanyahu, and Wolf 2017)

  • Aim to develop adaptive intelligent tutoring systems that provide personalized feedback and supplementary exercises to students who exhibit gaming behavior, thereby improving your learning outcomes and discouraging gaming practices. (NA?)

Healthcare And Medicine

  • Consider leveraging large language models and multi-prompt engineering with medical knowledge injection when conducting few-shot learning for chronic disease management, particularly in detecting mental disorders through user-generated textual content. (Haoxin Liu et al. 2024)

  • Consider combining brain-window theory with holographic learning to open up new possibilities for neurological diagnosis and the creation of novel fuzzy neural networks. (M. Yang, Huang, and Peng 2024)

  • Leverage the power of large language models, such as ChatGPT, to enhance medical imaging analysis by improving clinical workflow efficiency, reducing diagnostic errors, and assisting healthcare professionals in providing timely and accurate diagnoses. (M. Hu et al. 2023)

  • Consider utilizing ChatGPT as a dynamic honeypot interface in cyber security due to its ability to adapt to attackers actions, provide insights into your tactics, techniques, and procedures, and potentially delay or deter them from accessing critical network assets.’ (McKee and Noever 2023)

  • Focus on collecting and compiling medical dialogues in multiple languages, particularly Chinese, to effectively train and fine-tune large language models for improved precision and accessibility in the medical domain. (H. Xiong et al. 2023)

  • Carefully craft prompts for ChatGPT to effectively exploit its abilities in software vulnerability detection, considering factors such as structural and sequential code modeling, multi-round dialogues, and chain-of-thought reasoning. (Chenyuan Zhang et al. 2023)

  • Consider employing multiple adaptation strategies for large language models (LLMs) in biomedical and health applications, such as pre-training from scratch or checkpoints, fine-tuning with task-specific data, instruction fine-tuning and/or RLHF fine-tuning, soft prompt tuning, and prompt engineering, depending on the availability of data, computational resources, and expertise. (S. Tian et al. 2023)

  • Utilize ChatGPT in classrooms to stimulate students critical thinking skills by having them create, analyze, and revise ChatGPT outputs, thereby promoting awareness of the limitations of AI tools and appreciation for higher-order thinking skills.’ (Bitzenbauer 2023)

  • Focus on evaluating large language models (LLMs) on highly-specialized topics rather than widely available exams, as this allows for a more accurate assessment of your true potential. (Holmes et al. 2023)

  • Use interpretable graph learning techniques to develop accurate and transparent models for screening normal endoscopic large bowel biopsies, thereby significantly reducing pathologist workload and improving turnaround times. (S. Graham et al. 2023)

  • Prioritize the use of pretrained language models (LM) when extracting social determinants of health (SDOH) information from clinical notes, as they demonstrated superior performance in comparison to other methods. (Lybarger, Yetisgen, and Uzuner 2023)

  • Carefully evaluate the validity and applicability of artificial intelligence (AI) and machine learning-based interventions in clinical practice, considering factors such as bias, human values, regulatory compliance, and ethics, while developing robust standards for evaluating your effectiveness. (Haug and Drazen 2023)

  • Consider utilizing Electronic Health Record (EHR) foundation models like CLMBR to enhance the robustness of your clinical prediction models against temporal distribution shifts. (L. L. Guo et al. 2023)

  • Carefully evaluate the accuracy and completeness of ChatGPT-generated discharge summaries before using them in clinical settings, considering factors such as data privacy, patient acceptability, and the need for human oversight. (S. B. Patel and Lam 2023)

  • Consider conducting surveys to evaluate the effectiveness and user perception of AI-based chatbots in healthcare settings, particularly focusing on distinguishing between bot and human responses and measuring trust levels across varying degrees of health-related complexity. (Nov, Singh, and Mann 2023)

  • Consider using general-purpose AI language models, such as GPT-3, for healthcare applications, as they can achieve diagnostic accuracy close to that of physicians and better than lay individuals, although your triage performance remains lower. (Levine et al. 2023)

  • Carefully consider the limitations of language models like ChatGPT, such as misalignment with user intent, hallucination of information, and inability to attribute factual information to a source, when designing clinically-oriented prompts for use with them. (A. Rao, Kim, et al. 2023)

  • Focus on comprehensive clinical vignettes as a model for evaluating the performance of large language models (LLMs) like ChatGPT in the clinical decision-making process, considering your ability to integrate vast amounts of textual information and adapt to changing clinical scenarios. (A. Rao, Pang, et al. 2023)

  • Carefully examine the various stages of machine learning deployment workflow, including data management, model learning, model verification, and model deployment, to identify and address the numerous challenges faced by practitioners in each phase. (Paleyes, Urma, and Lawrence 2022)

  • Exercise caution when selecting training data for machine learning models, particularly in settings where there is a risk of spurious correlations, such as with medical imaging. (Berenguer et al. 2022)

  • Prioritize collecting and analyzing data from the same speakers with and without COVID-19 infection to minimize bias and improve the accuracy of speech-based COVID-19 detection algorithms. (Triantafyllopoulos et al. 2022)

  • Utilize artificial neural networks (ANNs) for concentration-time curve predictions due to your ability to accurately predict pharmacokinetics (PK) profiles without any predefined PK model, while being efficient in terms of training and prediction time. (Bräm et al. 2022)

  • Consider utilising ChatGPT, an AI language model, to enhance various aspects of your work, such as text summarisation, question answering, data collection, language translation, and writing assistance. However, they must critically evaluate the outputs produced by ChatGPT for accuracy and completeness before incorporating them into your research. (Moons et al. 2022)

  • Consider implementing a non-contact IoT-based system for real-time in-bed respiratory rate monitoring, which uses machine learning algorithms to analyze respiration-associated signals collected by contactless bed sensors, offering benefits such as low awareness of operation, upgradable RR estimation methods, scalable and extensible ecosystem, and remote monitoring capabilities. (Qingju Liu et al. 2021)

  • Employ machine learning techniques on digitized diplomatic documents to generate time-series data on elite threat perception, allowing for a more nuanced understanding of interstate dynamics. (Trubowitz and Watanabe 2021)

  • Focus on creating a comprehensive, realistic dataset for activity recognition of construction workers using IMU devices, addressing the unique challenges posed by construction sites and ensuring the dataset is applicable to real-world scenarios. (Mäkela et al. 2021)

  • Utilize multiple data sources and consider a broad range of variables when studying the COVID-19 pandemic, while acknowledging the limitations and challenges associated with the available data. (Alamo et al. 2020)

  • Consider using CLAM, a deep-learning-based weakly-supervised method that employs attention-based learning to automatically identify sub-regions of high diagnostic value in order to accurately classify the whole slide, while also utilizing instance-level clustering over the representative regions identified to constrain and refine the feature space. (M. Y. Lu et al. 2020)

  • Consider the security and robustness of machine learning (ML) / deep learning (DL) models in healthcare applications, as these models may be vulnerable to adversarial attacks and data poisoning, potentially compromising the accuracy and reliability of predictions. (Qayyum et al. 2020)

  • Carefully consider and address the 20 critical questions regarding transparency, reproducibility, ethics, and effectiveness (TREE) in your machine learning and artificial intelligence studies to improve the quality and clinical relevance of your research. (Vollmer et al. 2020)

  • Employ machine learning algorithms, particularly logistic regression and multinomial Naive Bayesian classifiers, to analyze clinical reports and accurately identify COVID-19, SARS, ARDS, and coexisting conditions, thereby enabling timely interventions and potentially saving lives. (Khanday et al. 2020)

  • Carefully consider and address potential biases in reported COVID-19 deaths, incorporate multiple data sources like cases and hospitalization rates as leading indicators, and apply appropriate statistical methods to estimate past and future deaths, infections, and hospitalizations. (“Modeling COVID-19 Scenarios for the United States” 2020)

  • Consider utilizing a coarse-to-fine generative adversarial network (GAN) for efficient data augmentation in the field of brain tumor segmentation, leading to improved performance in comparison to traditional augmentation approaches. (Mok and Chung 2019)

  • Focus on developing advanced machine learning algorithms and incorporating various types of features, including lexical and latent features, to improve the performance of case law retrieval and entailment tasks. (Kano et al. 2019)

  • Carefully consider the choice of activation functions when building deep learning models, as modern activation functions like ReLU and Leaky ReLU help overcome vanishing gradient issues and facilitate effective training of deeper networks. (Maier et al. 2018)

  • Consider utilizing deep feature representation learning on source code for automated software vulnerability detection, as demonstrated through the development of a fast and scalable vulnerability detection tool that achieved promising results on real software packages and the NIST SATE IV benchmark dataset. (Russell et al. 2018)

  • Consider integrating multilevel attention models within an end-to-end trainable CNN-RNN architecture to effectively highlight meaningful text words and image regions, thereby improving the accuracy of disease classification and report generation in chest X-ray analysis. (Xiaosong Wang et al. 2018)

  • Consider combining various data sources and employing advanced natural language processing (NLP) techniques to effectively capture the semantic and syntactic structures within electronic health record (EHR) narratives, enabling accurate computational phenotyping for diverse applications. (Zexian Zeng et al. 2018)

  • Consider utilizing a comprehensive machine learning approach for DNA methylation-based classification of central nervous system tumours, which can significantly enhance diagnostic precision and potentially revolutionize tumour pathology. (Capper et al. 2018)

  • Consider using a 3D deep convolutional neural network called “DeepNAT” for the automatic segmentation of neuroanatomy in T1-weighted magnetic resonance images, as it offers improved accuracy through end-to-end learning, multi-task learning, hierarchical segmentation, and incorporation of spectral coordinates. (Wachinger, Reuter, and Klein 2018)

  • Consider developing and implementing a predictive deep learning model for detecting critical radiological findings such as intracranial hemorrhages (ICH) in a quality improvement setting, potentially leading to reduced interpretation times and improved patient care. (Arbabshirani et al. 2018)

  • Leverage deep learning algorithms to analyze Electronic Health Records (EHRs) in your entirety, allowing them to identify complex patterns and relationships among various clinical parameters, thereby improving the accuracy and efficiency of predictive models in healthcare. (Rajkomar et al. 2018)

  • Consider employing a combination of traditional Generalized Linear Model (GLM) approaches alongside innovative Heteroscedastic Gaussian Process (hetGP) techniques to improve the accuracy and flexibility of your forecasting models, especially when dealing with complex and evolving phenomena like dengue fever incidence. (L. R. Johnson et al. 2018)

  • Employ a Random Forest machine learning algorithm for effective anomaly detection in IoT networks, achieving a high classification accuracy of 99.34% while minimizing false positives. (Angrishi 2017)

  • Conduct randomized clinical trials to evaluate the effectiveness of machine learning algorithms in improving patient outcomes, particularly in areas where early intervention is critical, such as severe sepsis. (Shimabukuro et al. 2017)

  • Consider utilizing deep convolutional neural networks (DCNN) for fault diagnosis in planetary gearboxes because they can learn features from raw data and optimize a combination of different fusion levels adaptively, leading to improved diagnostic accuracy. (Luyang Jing et al. 2017)

  • Consider implementing a “loopback” strategy in your deep learning models, which involves feeding the outputs of one part of the model back into another part, allowing for improved classification performance and increased interpretability. (Bastani, Kim, and Bastani 2017)

  • Consider employing a novel edge weight method for visibility graphs within complex networks to improve the accuracy of epilepsy detection from EEG signals. (Supriya, Siuly, and Zhang 2016)

  • Consider incorporating temporal relations when analyzing electronic health record (EHR) data, as doing so may improve model performance in predicting initial diagnosis of heart failure (HF) compared to conventional methods that ignore temporality. (E. Choi et al. 2016)

  • Develop a weakly-supervised multi-label image classification and disease localization framework to effectively analyze and interpret chest X-ray images, given the complexity and variation in size of pathological regions within the images. (Hariharan and Girshick 2016)

  • Utilize pluripotent stem (iPS) cells to create scalable sources for tissue-specific cell types, allowing them to develop reproducible 3D neural constructs that can effectively incorporate vascular and microglial components for developmental neurotoxicity screening purposes. (M. P. Schwartz et al. 2015)

  • Consider utilizing the scikit-learn Python library for neuroimaging data analysis, as it provides a comprehensive suite of machine learning algorithms that can effectively handle high-dimensional datasets, enabling accurate modeling of brain activities and behaviors. (Abraham et al. 2014)

  • Utilize machine learning techniques and natural language processing tools to analyze speech samples in order to accurately differentiate between control participants and participants with primary progressive aphasia (PPA) or semantic dementia (SD), as well as between the two patient groups. (Fraser et al. 2014)

  • Consider utilizing signal processing techniques, such as generating artificial EEG trials from the few EEG trials initially available, to augment the training set size and thereby reduce calibration time in oscillatory activity-based Brain-Computer Interfaces. (Congedo, Barachant, and Andreev 2013)

  • Use the newly introduced ARGO dataset, which contains 2034 anonymized recordings from nine post-ischemic VT patients, including 2.5 s bipolar EGMs, unipolar EGMs, the 12-leads surface ECGs, and the electroanatomic maps (voltage and local activation time maps), to develop and validate algorithms for detecting and delineating abnormal ventricular potentials (AV (Koplan and Stevenson 2009)

  • Utilize a combination of record filtering and nonparametric change-point detection tests to effectively analyze massive amounts of internet traffic data and accurately detect network anomalies in real-time. (Lévy-Leduc and Roueff 2009)

  • Focus on developing a unified framework for shilling attack detection in review systems for personalized recommendation, rather than creating separate algorithms for each specific attack strategy. (Burke et al. 2006)

  • Carefully consider and compare various sampling norms, including Bayesian diagnosticity, information gain, Kullback-Leibler distance, probability gain, and impact, when choosing the optimal method for evaluating the usefulness of questions in probabilistic evidence-gathering situations. (J. D. Nelson 2005)

  • Consider utilizing machine learning techniques, specifically genetic algorithms and decision trees, to automatically generate rules for classifying network connections in intrusion detection systems, thereby improving the accuracy and efficiency of identifying anomalous activities. (NA?)

  • Focus on creating cost-sensitive intrusion detection models that balance the trade-offs between various cost factors, including development cost, operational cost, damage cost due to successful intrusions, and the cost of manual and automated response to intrusions, in order to optimize the overall cost-effectiveness of intrusion detection systems. (NA?)

  • Carefully select the appropriate type of artificial neural network (ANN) for your specific application, considering factors such as computational efficiency, interpretability, and generalizability, while also taking into account potential limitations such as redundancy and overfitting. (NA?)

  • Utilise the Support Vector Machine (SVM) classification algorithm for structure-activity relationship analysis in drug design, as it outperforms various other machine learning techniques in a benchmark test. (NA?)

  • Carefully select and abstract features from high-dimensional, sparse, and noisy fMRI data to effectively train machine learning classifiers for decoding cognitive states. (NA?)

  • Utilize machine learning algorithms like boosted decision trees to effectively detect and classify malicious executables in the wild, achieving high accuracy rates while maintaining scalability. (NA?)

  • Consider using the LiveNet system, a flexible wearable platform for long-term ambulatory health monitoring with real-time data streaming and context classification, to develop distributed real-time multi-modal and context-aware applications for rehabilitation medicine. (NA?)

  • Use a combination of weakly orientation-selective fMRI voxels to accurately differentiate subtle variations in perceived stimulus orientation on a trial-by-trial basis, demonstrating a tight coupling between brain states and subjective mental states. (NA?)

  • Carefully choose relevant input variables for your machine learning models, considering factors like ecological significance, physical insights, and practical constraints of data collection, while also utilizing techniques like analyzing neural network weights and genetic programming to further refine the selection of significant input variables. (NA?)

  • Utilize machine learning-based classification techniques to proactively detect and identify botnet traffic, specifically focusing on distinguishing between IRC and non-IRC traffic, and then further refine the classification to separate botnet and real IRC traffic. (NA?)

  • Consider using semi-supervised learning algorithms like Co-Forest, which effectively integrates unlabelled data with limited labelled data, to build robust computer-aided diagnostic models. (NA?)

  • Explore and evaluate the performance of various machine learning paradigms, such as decision trees (DT), support vector machines (SVM), hybrid DT-SVM, and ensemble approaches, in the context of intrusion detection systems (IDS) to enhance the accuracy and robustness of these systems. (NA?)

  • Utilize shrinkage estimators to improve the accuracy of covariance matrix estimation in high dimensional sensor spaces, leading to better outcomes in single-trial ERP classification using linear discriminant analysis (LDA). (NA?)

  • Consider employing a combination of linguistic and biological annotation methods to improve the accuracy and reliability of your text mining analyses in the bio-medical domain. (NA?)

  • Consider using Support Vector Machines (SVMs) - a type of machine learning algorithm - when analyzing Magnetic Resonance (MR) imaging data for the early detection of individuals at risk of developing psychosis. (NA?)

  • Consider implementing a sophisticated adaptation scheme in BCI systems, which guides the user from an initial subject-independent classifier operating on simple features to a subject-optimized state-of-the-art classifier within one session, allowing for improved BCI performance and potentially solving the issue of BCI illiteracy. (NA?)

  • Employ a neurofeedback-based motor imagery training system for EEG-based brain-computer interface (BCI) applications, as it enhances classification accuracy and helps individuals better understand and execute motor imagery tasks. (NA?)

  • Consider implementing a fine-grained power monitoring system, such as ViridiScope, which utilizes ambient signals from inexpensive sensors placed near appliances to estimate power consumption, providing comprehensive coverage, fine-grained reporting, and easy installation. (NA?)

  • Carefully consider the unique challenges faced by anomaly detection systems in network intrusion detection, such as the high cost of errors, lack of training data, semantic gaps, and difficulties in evaluation, and adjust your methods accordingly to improve the effectiveness of machine learning techniques in this domain. (NA?)

  • Utilize data-driven methods like Random Forest (RF) analysis when developing cancer staging systems, which allows for more accurate predictions of patient outcomes by accounting for complex interactions among variables and nonlinear effects. (NA?)

  • Compare and rank various modeling techniques when predicting site index in Mediterranean mountain forests to ensure accurate predictions. (NA?)

  • Focus on improving the accuracy of identifying medication durations and reasons in clinical texts, as current state-of-the-art natural language processing systems struggle with these aspects. (NA?)

  • Consider combining the strengths of blacklist-based and feature-based methods in a unified framework to develop a robust anti-phishing solution that effectively balances true positive rates and false positive rates. (NA?)

  • Consider utilizing modified multiscale entropy (mMSE) as a potential biomarker for identifying infants at high risk for Autism Spectrum Disorder (ASD) based on resting state EEG data, given its demonstrated ability to differentiate typically developing children from high-risk groups with over 80% accuracy. (NA?)

  • Consider employing advanced data mining techniques such as Neural Networks, Support Vector Machines, Classification Trees, and Random Forests alongside traditional statistical methods like Linear Discriminant Analysis and Logistic Regression to enhance the predictive power of neuropsychological tests in identifying individuals with Mild Cognitive Impairment who are likely to progress to dementia. (NA?)

  • Utilise the NeuroSynth framework to enable large-scale automated neuroimaging meta-analyses, allowing for quantitative distinctions between forward and reverse inference, thereby improving the specificity and accuracy of mappings between neural and cognitive functions. (NA?)

  • Consider utilizing mobile health (mHealth) tools, specifically wearable sensors like accelerometers, gyroscopes, and pressure-sensitive textiles, to collect objective, continuous data on daily activities. This data can be processed using machine-learning algorithms to recognize movement patterns, allowing for improved monitoring of patients, more accurate outcome assessments in clinical trials, and increased understanding of the impact of various interventions on daily functioning. (NA?)

  • Consider utilizing a knife-edge shaped slit combined with a gamma camera for efficient and effective real-time monitoring of prompt gamma radiation emitted from the proton track in spot-scanning proton therapy, potentially improving dose delivery accuracy and reducing side effects. (NA?)

  • Consider using multiple spatial scales and various predictor variables when developing species distribution models, as this can improve the reliability and accuracy of the models. (NA?)

  • Consider combining multiple structural MRI analysis techniques to enhance classification accuracy in early detection of Alzheimers disease.’ (NA?)

  • Consider using multi-modal mobile sensing systems to simultaneously assess mental and physical health, as these systems can continuously capture fine-grained motion and privacy-sensitive audio data, allowing for derivation of metrics that reflect the results of commonly used surveys for assessing well-being by the medical community. (NA?)

  • Focus on developing efficient filters to quickly discard benign web pages, allowing for more accurate and timely analysis of potential malicious content. (NA?)

  • Consider employing Generalized Matrix Relevance LVQ (GMLVQ) for analyzing steroid metabolite excretion data, as it demonstrates promising results in detecting malignancy in adrenal tumors. (NA?)

  • Consider utilizing a One-Class Support Vector Machine (OCSVM) for classification tasks involving highly imbalanced datasets, particularly when dealing with malware detection on Android devices. (NA?)

  • Consider employing sensor-based energy modeling as a hybrid approach between forward’ and ‘inverse’ modeling, allowing the data to drive the model selection instead of relying solely on engineering domain knowledge.’ (NA?)

  • Avoid using non-causal methods when analyzing BCI competition data, as it can give an unfair advantage compared to real-world BCI practitioners. (NA?)

  • Consider combining multiple sources of information, such as chemical, biological, and phenotypic properties, when developing machine learning models for predicting adverse drug reactions (ADRs). (NA?)

  • Employ a Random Forests classification method, an ensemble classifier that randomly selects features for each individual classification tree, for robust detection of malicious PDF documents, achieving high true positive rates and low false positive rates even on previously unseen malware. (NA?)

  • Consider leveraging large datasets of ADOS evaluations and applying machine-learning techniques to develop efficient and accurate classifiers for autism diagnosis, potentially leading to earlier intervention and better patient outcomes. (NA?)

  • Consider employing a combination of Markov decision processes and dynamic decision networks to simulate clinical decision-making, allowing them to explore various healthcare policies, payment methodologies, and ultimately create a foundation for clinical artificial intelligence. (NA?)

  • Incorporate insights from instructional design literature to optimize the effectiveness of spontaneous BCI training protocols, focusing on improving feedback, instructions, and training tasks. (NA?)

  • Consider utilizing multivariate pattern analyses, particularly those based on machine learning models, to increase sensitivity in detecting spatially distributed effects in neuroimaging data, thereby allowing for a broader range of research questions to be explored. (NA?)

  • Carefully validate your classifiers using independent test sets, minimize the number of attributes in your models to prevent overfitting, and ensure the transportability of your methods across institutions. (NA?)

  • Utilize a combination of manual and semi-automated methods to filter and analyze social media data for adverse event detection, mapping colloquial language to a standard regulatory dictionary for accurate interpretation. (NA?)

  • Focus on developing a microwave-based system for pre-hospital stroke diagnosis, utilizing machine learning algorithms and incorporating leave-one-out validation methods combined with a Monte Carlo based bootstrap step to evaluate the detection methodology. (NA?)

  • Employ a multifaceted approach, integrating various domains like history, personality, and brain, to accurately predict and understand adolescent alcohol misuse. (NA?)

  • Consider using prompt gamma imaging (PGI) with a slit camera for accurate and real-time range verification in proton therapy, as it offers advantages over positron emission tomography (PET) and other indirect methods. (NA?)

  • Use cross-validation techniques to ensure independence of training and testing data, address potential biases caused by class imbalances, and carefully select appropriate evaluation metrics to accurately reflect the performance of your models. (NA?)

  • Utilise both structured and unstructured data from electronic medical records (EMRs) to develop accurate phenotype algorithms, incorporating natural language processing (NLP) techniques to extract relevant information from narrative clinical text. (NA?)

  • Develop a clear conceptual framework for integrating novel data streams (NDS) into public health surveillance, focusing on identifying opportunities for using NDS and applying minimal tests of validity and utility, while involving public health authorities and considering the diverse objectives and scales across various agencies. (NA?)

  • Adopt a standardized, automated, and non-controversial early-stage preprocessing pipeline (the PREP pipeline) for EEG data to ensure accurate and reliable downstream analysis across multiple collections. (NA?)

  • Carefully evaluate and compare different feature selection and classification methods when developing radiomics-based prognostic models for head and neck cancer, considering both prognostic power and stability. (NA?)

  • Focus on selecting appropriate feature selection and classification methods, such as the Wilcoxon test based feature selection method WLCX and/or random forest (RF) classification method, to achieve high performance and reasonable stability in radiomics based predictive studies. (NA?)

  • Avoid relying solely on theoretical chance levels when evaluating decoding accuracy in brain signal classification studies, especially when dealing with small sample sizes, and instead utilize statistical approaches like binomial cumulative distributions or permutation tests to ensure accurate and meaningful interpretation of results. (NA?)

  • Use an ex vivo platform that accurately reflects tumor heterogeneity to improve prediction of clinical responses to anticancer drugs. (NA?)

  • Consider conducting dense longitudinal phenotyping studies to capture the dynamics of brain function over extended periods, which can help improve understanding of major psychiatric disorders. (NA?)

  • Utilize advanced image analysis and machine learning algorithms to identify complex and reproducible imaging patterns predictive of overall survival and molecular subtype in glioblastoma (GB). (NA?)

  • Consider combining multiple data sources and applying machine learning techniques to improve the accuracy and reliability of your predictions, particularly in areas like influenza surveillance. (NA?)

  • Consider using computer-based models to analyze large amounts of digital footprint data, as these models can often produce more accurate and reliable personality judgments than traditional human evaluators. (NA?)

  • Use multiple feature selection and classification methods to increase the robustness and generalizability of radiomics-based predictive models, particularly considering the “curse of dimensionality” inherent in high-throughput data-mining fields. (NA?)

  • Use consensus clustering to reduce feature redundancy in radiomics, allowing them to identify and validate non-redundant sets of imaging biomarkers, ultimately leading to improved prognostic performance in lung and head & neck cancer. (NA?)

  • Utilize a global vis-NIR spectral library to analyze soil characteristics and variations across different regions, enabling accurate predictions of soil attributes and conditions through advanced statistical techniques and machine learning algorithms. (NA?)

  • Consider using machine learning techniques, specifically ensemble methods like “boosting”, to analyze complex datasets with multiple variables and interactions, as demonstrated by the studys successful prediction of 5-year all-cause mortality in patients with suspected coronary artery disease.’ (NA?)

  • Consider creating and utilizing large, diverse, and well-curated datasets like the TUH-EEG Corpus to effectively train machine learning algorithms and advance biomedical research. (NA?)

  • Use a combination of correlation-based feature elimination and univariate feature selection to reduce redundancy in radiomic feature space, followed by application of various machine learning classification methods to accurately predict lung cancer histologic subtypes. (NA?)

  • Utilize large-scale, high-quality datasets, employ robust statistical techniques, and incorporate multiple omics layers to enhance the accuracy and reliability of your findings in health research. (NA?)

  • Consider employing a multi-scale Convolutional Neural Network (CNN) for accurate and consistent segmentation of Magnetic Resonance Imaging (MRI) brain images across different age groups and orientations, without explicitly defining spatial features. (NA?)

  • Consider employing Deep Learning (DL) techniques, specifically Deep Neural Networks (DNNs), for the classification of drugs into therapeutic categories based on your transcriptional profiles. This approach demonstrated superior performance compared to traditional statistical methods such as Support Vector Machines (SVMs), particularly when using pathway activation scores as features. (NA?)

  • Utilise both data-driven and theory-driven approaches to overcome the challenge of high dimensionality in psychiatric data sets, while carefully considering potential issues like overfitting and ensuring appropriate validation procedures. (NA?)

  • Consider using machine-learning models to predict neurological impairment in individual subjects, taking into account both functional connectivity and lesion topography, as they may offer complementary insights into the underlying causes of various behavioral deficits. (NA?)

  • Utilise a comprehensive evaluation framework consisting of five components - overlap with CGC, agreement between methods, comparison of observed vs. theoretical P values, number of significant genes predicted, and prediction consistency on independent partitions of the dataset - to effectively evaluate and compare various driver gene prediction methods. (NA?)

  • Consider utilising multiple machine learning models, such as Boosted Regression Trees (BRT), Multivariate Discriminant Analysis (MDA), and Support Vector Machine (SVM), in conjunction with ensemble modelling to enhance the accuracy of predictive models in groundwater pollution risk assessments. (NA?)

  • Ensure transparency, informed consent, and proper regulations when dealing with sensitive medical data, especially when collaborating with private entities like Google DeepMind. (NA?)

  • Consider implementing multi-source transfer learning with convolutional neural networks for enhanced lung pattern analysis, particularly when dealing with limited medical imaging data. (NA?)

  • Utilize a combination of human effort and machine automation in order to optimize the efficiency and sustainability of living systematic reviews. (NA?)

  • Utilize a combination of a segmentation model trained on 3D annotated data and label propagation to achieve superior results in the segmentation of the Inferior Alveolar Nerve (IAN) in CBCT scans. (Kroon, n.d.)

  • Consider employing a multi-modal approach, incorporating diverse tasks and audio signals, to improve the accuracy of smartphone-based lung health assessments. (NA?)

  • Utilize deep learning techniques, particularly convolutional neural networks (CNNs), for improved accuracy and efficiency in brain MRI segmentation tasks, leveraging large datasets and taking advantage of advances in GPU processing power. (NA?)

  • Consider utilizing smartphone applications for collecting real-time, objective data on physical activity, fitness, and sleep, as it demonstrates feasibility and potential benefits for large-scale population health studies. (NA?)

  • Consider employing machine learning techniques to detect ransomware based on your power usage patterns on IoT nodes, particularly Android devices, as this approach increases the detection rate to 95.65%, making it a potentially valuable addition to existing malware detection strategies. (NA?)

  • Consider both with-contact and contactless remote health monitoring systems when conducting studies, as both types of systems play crucial roles in providing accurate and reliable physiological data for patients. (NA?)

  • Utilise a combination of local and global contextual features in your Convolutional Neural Networks (CNNs) for optimal brain tumor segmentation. (NA?)

  • Consider utilizing deep learning techniques, specifically convolutional neural networks (CNNs), for medical image analysis tasks across various domains, as they have demonstrated promising results and efficacy in numerous studies. (NA?)

  • Combine distinct measurements of biological aging, such as neuroimaging and epigenetics, to better determine risk of age-related deterioration and death. (NA?)

  • Avoid assuming that the computational complexity of machine-learning potentials (MLPs) must grow with the number of chemical species involved, as it demonstrates that the same model complexity that works optimally for a ternary material can also effectively describe a system with 11 chemical species. (NA?)

  • Adopt a consistent reference frame for method benchmarking to advance the field of MRI-based radiation therapy. (NA?)

  • Utilize the Open University Learning Analytics Dataset (OULAD) for evaluating predictive models, comparing models developed by other researchers, and analyzing the influence of virtual learning environments on learning outcomes. (NA?)

  • Use a multi-stream multi-scale convolutional network architecture to improve the accuracy of automatic nodule classification in lung cancer screening, as it outperforms traditional machine learning methods and aligns closely with the inter-observer variability among experienced human observers. (NA?)

  • Carefully consider the limitations of electronic medical records (EMRs) when analyzing and interpreting the data, particularly regarding the identification of disease and symptom mentions, the complexity of real-world patient cases, the potential confusion between correlation and causation, and the impact of physician decision-making on the recording of observations. (NA?)

  • Consider integrating multiple omics data sources (such as genomics, transcriptomics, proteomics, metabolomics, etc.), rather than focusing solely on genomics, to effectively uncover underlying biological mechanisms and improve precision oncology. (NA?)

  • Consider employing a combination of traditional program analysis and deep learning approaches to create robust representations of PowerShell scripts, specifically through converting them to your Abstract Syntax Tree (AST) counterparts and building embedding vector representations of each AST node type based on a corpus of PowerShell programs. (NA?)

  • Utilize supervised machine learning techniques to estimate household characteristics from electricity consumption data, enabling personalized and scalable energy efficiency programs for private households. (NA?)

  • Employ machine learning techniques like Random Forest Models to effectively calibrate low-cost air quality sensors, thereby improving your accuracy and precision, and making them suitable for widespread deployment in air quality monitoring networks. (NA?)

  • Utilize artificial intelligence technologies, particularly deep learning, to build computer-aided detection (CAD) systems for chest radiography, as these methods have proven to surpass the performance of medical professionals in various medical imaging tasks. (NA?)

  • Carefully choose your clinical endpoints, appropriately measure model performance, increase model stability, and strive for interpretability in order to overcome the challenges faced in implementing machine learning-based solutions in personalized medicine. (NA?)

  • Consider using a deep neural network model with a large feature set extracted from electronic health records to predict patient mortality within 12 months, as a proxy for identifying patients who might benefit from palliative care. (NA?)

  • Consider utilizing machine learning algorithms like InSight to improve the accuracy of sepsis detection and prediction, especially in comparison to traditional rule-based disease-severity scoring systems like SIRS, MEWS, and SOFA. (NA?)

  • Focus on developing artificial intelligence (AI) models for well-defined tasks in medical settings, ensuring they complement rather than replace human doctors, while addressing concerns about trust and transparency. (NA?)

  • Use the IDRiD dataset, a comprehensive collection of high-quality retinal fundus images with precise pixel-level annotations of diabetic retinopathy lesions, optic disc, and fovea center coordinates, along with image-level disease severity grades, to train, validate, compare, and improve deep learning models for automated detection and classification of diabetic retinopathy and diabetic macular edema. (NA?)

  • Consider employing machine learning techniques like random forest models when dealing with complex datasets involving numerous predictors, especially if traditional regression models fail to adequately explain the observed variance. (NA?)

  • Utilize semi-parametric neural networks (SNN) for enhanced predictive performance in crop yield modeling, especially when dealing with complex nonlinear relationships in high-dimensional datasets, as demonstrated by the superior performance of SNN over traditional statistical methods and fully-nonparametric neural networks in predicting corn yields. (NA?)

  • Focus on developing and optimizing iterative reconstruction (IR) algorithms for CT imaging to effectively balance image quality and radiation dose reduction, taking advantage of recent technological advancements like GPU acceleration and incorporating prior knowledge to improve diagnostic accuracy. (NA?)

  • Carefully evaluate the benefits and challenges of integrating artificial intelligence (AI) into medical imaging, taking into account factors such as accuracy, reliability, ethical concerns, and the potential impact on the roles and responsibilities of radiologists. (NA?)

  • Consider utilising convolutional neural networks (CNNs) for automating the detection of gastric cancer in endoscopic images, given your demonstrated effectiveness in achieving high levels of diagnostic accuracy. (NA?)

  • Focus on developing AI-enabled healthcare systems that balance the limitations and promises of AI, integrating it into specific areas such as patient administration, clinical decision support, patient monitoring, and healthcare interventions while considering ethical, legal, and social implications. (NA?)

  • Consider employing multiple machine learning algorithms, feature selection techniques, cross-validation methods, and performance evaluation metrics to develop a robust and accurate heart disease prediction system. (NA?)

  • Use a combination of machine learning techniques and density functional theory calculations to rapidly screen hybrid organic-inorganic perovskites (HOIPs) based on bandgap, solving the problems of toxicity and poor environmental stability in HOIPs. (NA?)

  • Focus on training a decoder on single concepts that comprehensively cover the semantic space, leading to the ability to robustly decode meanings of semantically diverse new sentences. (NA?)

  • Consider utilizing machine learning methods to analyze extensive databases containing diverse materials properties, enabling the prediction of superconducting critical temperatures and identification of potential new superconductors. (NA?)

  • Prioritize patient care when considering the implementation of AI in healthcare, actively participate in the evolution of AI technologies, and invest in improving your own human abilities such as empathy, communication, and decision-making. (NA?)

  • Carefully evaluate the validity and fairness of predictive algorithms like COMPAS, especially when they are used in consequential domains like criminal justice, as simpler models with fewer features can often achieve comparable accuracy while avoiding potential biases. (NA?)

  • Utilise high-resolution magnetic resonance imaging (MRI) and advanced data pre-processing techniques, particularly non-linear image alignment, to investigate the contribution of subcortical brain nuclei to human cognition and behaviour. (NA?)

  • Utilise nonlinear analysis of EEG signals alongside pattern classification techniques to identify early indicators of autism spectrum disorder (ASD) in infants as young as three months old. (NA?)

  • Utilise deep neural networks for automatic vetting of Kepler Transit Candidates (TCEs), as these models can effectively distinguish between genuine transit signals and false positives, thereby improving the efficiency and accuracy of exoplanet discovery. (NA?)

  • Embrace the opportunities provided by big data, including the use of naturalistic and crowdsourced data, methodological advancements, and the creation of new data resources, to advance psychological theory and enhance the generalizability of findings beyond traditional laboratory settings. (NA?)

  • Consider utilizing deep learning techniques, specifically convolutional neural networks (CNNs), for improved cancer diagnosis accuracy, leveraging publicly available datasets for testing and validation purposes. (NA?)

  • Consider employing advanced machine learning techniques, particularly deep learning, to analyze big data sets in order to discover complex relationships between molecular representations and observed phenomena, thereby producing more accurate and generalizable insights for drug discovery. (NA?)

  • Consider applying advanced machine learning techniques such as lasso regression, random forest, gradient boosted decision trees, and deep neural networks when developing triage systems for emergency departments, as these models can significantly outperform traditional methods like the Emergency Severity Index (ESI) in terms of accuracy and efficiency. (NA?)

  • Carefully consider the ethical implications and potential biases when using artificial intelligence (AI) tools to study mental health and illness. (NA?)

  • Carefully consider the choice of deep learning models for network traffic classification based on the selected features, as this directly impacts input structure, computational complexity, and memory complexity. (NA?)

  • Use a hierarchical recurrent neural network called SeqSleepNet for sequence-to-sequence automatic sleep staging, which enables accurate classification of sleep stages across multiple epochs simultaneously. (NA?)

  • Consider utilizing a diverse range of deep learning techniques, such as convolutional neural networks, recurrent neural networks, and generative adversarial networks, to effectively tackle various cybersecurity challenges like malware detection, spam filtering, and network intrusion detection. (NA?)

  • Consider integrating machine learning techniques into your causal inference processes, particularly during the pre-final estimation stages, to enhance the accuracy and efficiency of your results. (NA?)

  • Utilise a combination of deep learning and handcrafted feature extraction techniques to improve the classification accuracy of lung abnormalities in chest x-ray and lung CT scan images. (NA?)

  • Carefully address the black box’ issue in AI algorithms, ensuring transparency and explainability in order to build trust among users and mitigate potential risks associated with opaque decision-making processes.’ (NA?)

  • Prioritize improving the progression of clinical NLP methods from extraction towards understanding, recognizing relations among entities instead of entities in isolation, temporal extraction to understand past, present, and future clinical events, exploiting alternative sources of clinical knowledge, and ensuring the availability of large-scale, de-identified clinical corpora. (NA?)

  • Consider using artificial intelligence (AI) systems for breast cancer screening due to your ability to outperform human radiologists in terms of specificity and sensitivity, thus improving overall accuracy and efficiency. (NA?)

  • Conduct external validation of AI algorithms for diagnostic analysis of medical images using adequately sized datasets collected from multiple institutions in a prospective manner, ensuring they accurately represent the manifestation spectrum of target patients in real-world clinical settings. (NA?)

  • Shift your analytical focus from group means to understanding cohort variation, moving from first- to second-order statistics, and mapping deviations at the level of individual. (NA?)

  • Critically evaluate the methods used in prior studies before attempting to replicate them, especially when dealing with complex datasets like those involving resting state fMRI and clinical data. (NA?)

  • Consider integrating machine learning and multiscale modeling to create robust predictive models that incorporate underlying physics to handle ill-posed problems and explore massive design spaces, thereby offering new insights into disease mechanisms, identifying new therapeutic targets and treatment strategies, and informing decision-making for improved human health. (NA?)

  • Consider combining photoplethysmography (PPG) data with other physiological signals like electrocardiograms (ECG), ballistocardiograms (BCG), and phonocardiograms (PCG) to improve the accuracy of blood pressure estimation. (NA?)

  • Focus on the initial human-AI onboarding phase, when users are first being introduced to an AI system, learning its capabilities, and determining how they will partner with it in practice, as this stage can significantly influence initial impressions, mental models, and strategies of use. (NA?)

  • Consider developing deep learning architectures specifically tailored to your specific medical imaging task, reducing computational and memory requirements without compromising classification performance, and utilizing visualization techniques such as saliency maps and grad-CAMs to enhance interpretability and potentially aid clinician decision-making. (NA?)

  • Consider integrating Artificial Intelligence (AI) and Machine Learning (ML) techniques into various stages of clinical trial design, specifically focusing on improving patient cohort composition, facilitating patient recruitment, and enhancing patient monitoring. (NA?)

  • Utilize machine learning algorithms, specifically Random Forest and Three-Way Random Forest, to analyze routine blood exam data for accurate detection of COVID-19 infections, offering a potentially faster, less expensive, and more accessible alternative to the current gold standard test, rRT-PCR. (NA?)

  • Consider utilizing deep learning algorithms, specifically the Truncated Inception Net, for accurate and efficient identification of COVID-19 positive cases in chest x-rays, achieving impressive accuracies of up to 99.96%. (NA?)

  • Carefully consider the trade-off between data privacy and public health while developing AI solutions for COVID-19, ensuring transparency and ethical practices in data collection and analysis. (NA?)

  • Consider employing advanced machine learning techniques like genetic algorithms, particle swarm optimization, and grey wolf optimization to improve the accuracy and generalizability of outbreak prediction models, particularly for complex scenarios like COVID-19. (NA?)

  • Employ deep learning-based CNN models like Xception, Inception V3, and ResNeXt to accurately classify chest X-ray images for COVID-19 detection, achieving the highest accuracy of 97.97% with the Xception model. (NA?)

  • Consider using a combination of record-wise and subject-wise cross-validation methods along with an augmented image database to improve the generalization capability and accuracy of your convolutional neural network model for brain tumor classification. (NA?)

  • Adopt a multidisciplinary perspective when studying explainability in medical AI, considering the various dimensions including technological, legal, medical, and patient perspectives. (NA?)

  • Utilize machine learning techniques to effectively analyze large-scale, heterogeneous datasets in order to uncover useful patterns that would be difficult or impossible for even well-trained individuals to identify, thereby revolutionizing the fields of clinical diagnostics, precision treatments, and health monitoring. (NA?)

  • Consider adopting a macroscale perspective on cortical organization to understand how the integrated nature of neural processing gives rise to function and dysfunction, utilizing tools like BrainSpace to analyze neural manifolds in a compact and reproducible manner. (NA?)

  • Consider employing a deep learning-based dynamic analysis system like DL-Droid for Android malware detection, particularly when using stateful input generation for enhanced code coverage and improved performance. (NA?)

  • Focus on collecting diverse training data rather than solely focusing on improving model architecture, as data diversity plays a significant role in enhancing the accuracy and reliability of lung segmentation algorithms. (NA?)

  • Carefully select the privacy budget when evaluating differential privacy mechanisms for pharmacogenetic models, as higher privacy budgets can increase the risk of stroke, bleeding events, and mortality. (NA?)

  • Utilize a mobile phone-based online survey combined with an AI algorithm to quickly identify potential COVID-19 cases, allowing for earlier isolation and reducing the chances of virus transmission. (NA?)

  • Follow the Image Biomarker Standardization Initiative (IBSI) guidelines for standardized feature calculations from all radiomic feature matrices, and ensure rigorous feature selection and dimension reduction procedures to minimize model overfitting. (NA?)

  • Adopt a rigorous approach to evaluating the acceptability, safety, and effectiveness of diverse health care conversational agents, particularly those driven by artificial intelligence and delivered via smartphone apps. (NA?)

  • Utilize active learning-based cross-population train/test models that incorporate multitudinal and multimodal data to effectively analyze and predict the spread of COVID-19. (NA?)

  • Adopt a multi-scale approach to applying artificial intelligence (AI) in combatting COVID-19, focusing on molecular, clinical, and societal scales, while ensuring adherence to regulatory and quality assurance frameworks and fostering international collaboration through multidisciplinary research and open science. (NA?)

  • Consider using counterfactual methods instead of traditional associative methods in order to improve the accuracy of medical diagnosis by better accounting for causal relationships among variables. (NA?)

  • Consider the potential limitations of deep learning models in accurately predicting phenotypes from structural and functional MRI data, particularly regarding the presence of nonlinear structures requiring both compositionality and translational invariance. (NA?)

  • Utilize Gaussian Process Regression (GPR) to accurately estimate the capacity and predict Remaining Useful Life (RUL) of Lithium-ion batteries using Electrochemical Impedance Spectroscopy (EIS) data, which contains richer information about battery health compared to signals currently tracked in battery management systems. (NA?)

  • Focus on developing and validating reliable biomarkers for pain, which would help in patient stratification, personalizing treatment plans, reducing variability in clinical trials, and providing objective measures of pain. (NA?)

  • Focus on developing and applying convolutional neural network (CNN)-based techniques for machine fault diagnosis, as they have demonstrated superior performance compared to traditional machine learning models and other deep learning architectures. (NA?)

  • Ensure that your work is transparent and reproducible, particularly when dealing with complex and rapidly evolving areas such as artificial intelligence and machine learning in medical applications. (NA?)

  • Carefully consider the type of CDSS being studied, its context of use, and the potential impact on both clinicians and patients, while employing robust experimental designs and rigorous analytical approaches to draw meaningful conclusions about the effectiveness and safety of these systems. (NA?)

  • Consider using transfer learning methods (pre-trained CNN weights on ImageNet) over models trained from scratch for improved performance in binary classification tasks like distinguishing normal vs. abnormal chest X-rays. (NA?)

  • Consider using a deep learning-based Convolutional Neural Network (CNN) model called Truncated Inception Net for screening COVID-19 positive chest X-rays (CXRs) from other non-COVID and/or healthy cases, achieving high accuracies of 99.96% and 99.92%, respectively. (NA?)

  • Utilise machine learning techniques to analyse large amounts of COVID-19 related clinical data, facilitated through the creation of a centralised database, to enable more accurate predictions, diagnoses, and therapies. (NA?)

  • Utilise the UCDP Candidate dataset alongside the UCDP GED dataset to enhance the timeliness and accuracy of conflict predictions. (NA?)

  • Utilize both stochastic theory mathematical models and data science/machine learning techniques for accurate COVID-19 forecasts, incorporating diverse data sources including big data from WHO/National databases and social media communications, while considering a wide range of relevant parameters such as environmental factors, incubation period, quarantine effects, age, gender, and others. (NA?)

  • Consider developing a second-by-second sleep apnea detection method using a 1-dimensional convolutional neural network (1D-CNN) for feature extraction and detection, which achieves a high resolution and outperforms several lower resolution state-of-the-art apnea detection methods. (NA?)

  • Consider employing machine learning techniques, particularly supervised learning, when studying COVID-19 cases, as it demonstrates improved accuracy rates compared to traditional methods. (NA?)

  • Adopt a structured literature review (SLR) method for studying the scientific corpus of a research field, enabling them to analyze both qualitative and quantitative variables, ensuring scientific rigor, reliability, and replicability of operations. (NA?)

  • Explore the potential benefits of integrating machine learning and deep learning algorithms into vibration-based structural damage detection systems, as these advanced technologies offer improved accuracy and robustness compared to traditional methods. (NA?)

  • Consider utilizing diverse deep learning approaches, such as supervised, weakly supervised, unsupervised, transfer learning, and various sub-variants, when conducting histopathological image analysis, taking into account the specific requirements and characteristics of the study. (NA?)

  • Consider combining deep-learning algorithms with natural-language models to decode words and sentences from cortical activity in individuals with speech disorders, potentially improving your communication abilities. (NA?)

  • Develop artificial intelligence-specific EQUATOR guidelines, especially STARD, to ensure accurate and consistent reporting of diagnostic accuracy in medical imaging studies. (NA?)

  • Leverage deep learning techniques, specifically convolutional neural networks, to achieve state-of-the-art performance in medical computer vision tasks, while addressing the unique challenges posed by medical data, such as limited availability and variability in quality. (NA?)

  • Consider adopting the Med-BERT approach when working with structured Electronic Health Records (EHRs) for disease prediction tasks, as it enables significant improvements in prediction accuracy by generating contextualized embeddings pretrained on large-scale EHR datasets. (NA?)

  • Carefully account for potential biases in your datasets, particularly when dealing with self-reported symptoms, and strive to use unbiased features whenever possible to improve the accuracy and reliability of your findings. (NA?)

  • Consider using domain-specific training materials when developing large language models for specific fields like ophthalmology, and incorporating measures of epistemic uncertainty into the models output to avoid misleading users.’ (NA?)

  • Carefully balance the tradeoffs between detector latency and specificity when developing algorithms for seizure detection, particularly due to the rarity of seizure events and limited availability of seizure training data. (NA?)

  • Consider utilizing domain adaptation techniques when dealing with the domain shift problem in medical image analysis, as it has proven effective in addressing the issue of differing distributions between source/reference data and target data. (NA?)

  • Consider adopting a Multi-modal Embedding Open Learner Model (MeOLM) framework, which integrates course embeddings, OLM embeddings, multi-modal embedding module, and task-specific modules, working together with GPT to enhance personalized learning. (NA?)

  • Carefully consider the ethical and legal implications of integrating AI into dental education, while ensuring that dental curricula are updated to include AI literacy and competencies for future dental professionals. (NA?)

  • Utilize large language models (LLMs) for message generation in health communication, as demonstrated by its successful application in generating folic acid awareness messages that outperformed human-generated counterparts in terms of quality and clarity. (NA?)

  • Utilise AI-generated suggestions alongside human-generated ones to enhance the efficiency and effectiveness of clinical decision support (CDS) alerts. (NA?)

  • Consider using the LiveNet system, a flexible wearable platform for long-term ambulatory health monitoring with real-time data streaming and context classification, to develop personalized, data-rich health profiles of users over time. (NA?)

  • Carefully consider the role of prompt engineering in mediating the effects of ChatGPT on students academic performance, as it plays a crucial part in optimizing the benefits of AI in educational settings.’ (NA?)

  • Adopt a “pragmatic” approach to developing a core dataset for multiple sclerosis (MS) by including clinically relevant variables that are feasible to collect, while acknowledging the limitations of this approach and planning for future revisions. (NA?)

  • Carefully evaluate the potential risks and benefits of implementing large language models (LLMs) in healthcare settings, advocating for the establishment of a comprehensive regulatory framework that addresses the unique characteristics of LLMs, such as your scale, complexity, hardware requirements, broad applicability, real-time adaptation, societal impact, and data privacy and security concerns. (NA?)

  • Develop a Medical Knowledge-enhanced Prompt Learning (MedKPL) model for diagnosis classification, which can effectively integrate medical knowledge into the models to enhance diagnosis and transfer learned diagnosis capacity to unseen diseases using alternating relevant disease knowledge. (NA?)

  • Consider the importance of diverse patient populations and global clinical trials when developing and evaluating AI/ML-enabled medical devices, given the observed limitations in the current landscape. (NA?)

  • Consider leveraging automatically derived samples from large amounts of social media data to study mental health disorders, rather than relying solely on surveys, as this approach allows for larger and more diverse datasets. (NA?)

Finance And Economics

  • Pay attention to the increasing complexity of bidding problems in real-time bidding display advertising, and adapt your optimization techniques accordingly, especially considering the growing importance of reinforcement learning methods. (Ou et al. 2023)

  • Focus on creating a larger dataset, extracting multiple static features from different angles, and utilizing advanced machine learning techniques to improve the accuracy and robustness of identifying smart Ponzi schemes in Ethereum. (Zibin Zheng et al. 2023)

  • Consider using semi-supervised graph learning techniques when analyzing financial transaction graphs to improve the accuracy of detecting potential money laundering activities. (Karim et al. 2023)

  • Consider using a neural network-based approach, like the proposed AMAP framework, to effectively detect and discover money laundering sub-networks within massive transaction networks, thereby providing a more comprehensive risk coverage and deeper insights into money laundering strategies. (Z. Chai et al. 2023)

  • Consider employing a probabilistic model for the quickest detection of credit card fraud, where for each transaction the posterior probability of being fraudulent is returned and a personalized threshold for each cardholder is optimally determined. (Buonaguidi et al. 2022)

  • Consider using AUC-oriented Graph Neural Networks (AO-GNN) when dealing with imbalanced labels in fraud detection tasks, as it effectively addresses the issue of noisy topological structures caused by fraudsters through a combination of classifier parameter searching and edge pruning policy searching. (Mengda Huang et al. 2022)

  • Combine the strengths of the Implicit Function Theorem (IFT) and Monotone Comparative Statics (MCS) approaches to effectively conduct comparative statics analysis in complex joint pricing and inventory management models. (N. Yang and Zhang 2022)

  • Carefully select training samples with larger price changes to improve the signal-to-noise ratio in the training data, allowing reinforcement learning agents to better identify regularities and create profitable trading strategies in high-frequency trading. (Briola et al. 2021)

  • Consider incorporating explainability into your fraud detection frameworks, specifically by integrating a hybrid explainer that combines task-aware measures of predictions generated by the GNNExplainer and task-agnostic centrality measures, allowing for a more comprehensive understanding of complex fraudulent patterns. (S. X. Rao et al. 2021)

  • Consider both model design and system design simultaneously, allowing for flexibility in trading off model performance against computational power costs. (Zhe Wang et al. 2020)

  • Focus on optimizing runtime instrumentation to minimize the number of instrumented load/store statements, as this reduces the overhead of runtime validation for smart contracts. (A. Li, Choi, and Long 2020)

  • Consider implementing a decentralized exchange (DEX) like SPEEDEX, which utilizes an Arrow-Debreu exchange market structure to achieve scalability, eliminate internal arbitrage opportunities, and prevent certain front-running attacks, ultimately providing a fair and efficient platform for secure asset trading. (Daian et al. 2019)

  • Consider utilizing a combination of graph-based detection and time series-based detection modules to improve the effectiveness and scalability of fraud detection systems in large-scale e-commerce platforms. (H. Weng et al. 2018)

  • Carefully consider the potential impact of your study on the broader scientific community and society, taking into account ethical considerations and the potential consequences of your findings. (Klarman, Flores, and Kuzmanovic 2018)

  • Utilise a combination of clustering heuristics, machine learning-based validation methods, and ground truth datasets to accurately track and analyse Bitcoin transactions, thereby improving understanding of the limitations of anonymity in these systems. (Goldfeder et al. 2017)

  • Consider incorporating risk-averse bidding strategies in your studies, which involve penalizing bids with high uncertain CTR (click-through rate) and rewarding those with greater confidence, leading to improved overall campaign profits. (Haifeng Zhang et al. 2017)

  • Integrate attribution modeling into your bidding strategy to enhance the efficiency of the bidding policy in the context of performance advertising. (Diemert et al. 2017)

  • Utilise Lagrangian relaxation techniques to transform complex, nonconvex problems into simpler, convex ones, allowing for efficient computation and improved accuracy in solving them. (Grigas et al. 2017)

  • Apply and compare the performance of multiple machine learning methods in familiar empirical problems to understand your usefulness in the field of asset pricing. (J. B. Heaton, Polson, and Witte 2016)

  • Consider the impact of downstream auctions on optimal reserve pricing strategies, as these can lead to improved revenue generation compared to traditional approaches. (Lisbona, Chammas, and Lee 2016)

  • Consider the complexities of online video advertising, including the differences in metrics and optimization goals compared to traditional display advertising, and develop tailored solutions accordingly. (Geyik et al. 2016)

  • Integrate the user response prediction and bidding optimization processes into a unified framework, allowing for simultaneous optimization of both components. (K. Ren et al. 2016)

  • Consider implementing a feedback control mechanism within your RTB system to optimize campaign-level performance by dynamically adjusting bids based on the deviation from the reference eCPC, ultimately leading to improved click numbers and lower costs. (Weinan Zhang et al. 2016)

  • Consider integrating multimodal data sources, specifically combining verbal and vocal cues from earnings calls, with graph-based techniques to account for stock interdependencies when developing models for stock volatility prediction. (Dhingra et al. 2016)

  • Employ a combination of network analysis and supervised learning to effectively analyze group behavior in financial transaction networks, thereby improving the accuracy of money laundering detection systems. (Savage et al. 2016)

  • Carefully consider the choice of loss function when developing models for online advertising auction bidding, as incorporating cost-sensitivity via weighted log loss (WNLL) can significantly improve both offline and online performance compared to traditional log loss (NLL) approaches. (Vasile, Lefortier, and Chapelle 2016)

  • Develop an efficient anomaly detection system for monitoring the performance of a large-scale demand side platform (DSP) by applying filtering and aggregation to millions of metrics, generating time series from the aggregated metrics, and deploying a simple algorithm to identify under-performing metrics. (B. Zhou and Shariat 2016)

  • Consider combining traditional DSGE model elements with unique aspects of the studied economy, such as fiscal policies, regulated pricing, external financing, and imported goods usage, while utilizing Bayesian estimation methods to improve parameter identification and incorporate expert knowledge. (Castro et al. 2015)

  • Develop a principled mathematical formulation and novel computational solution to mine and exploit arbitrage opportunities in real-time display advertising, thereby improving the efficiency and transparency of ad markets. (Weinan Zhang and Wang 2015)

  • Use a combination of online and offline data to develop a control-based method for optimizing budget pacing and campaign performance simultaneously in the context of digital advertising. (Jian Xu et al. 2015)

  • Utilise a data-driven approach to multi-touch attribution modelling in online advertising, specifically proposing a novel AdditiveHazard model based on survival theory. This model offers advantages over traditional rule-based models by allowing for the removal of presentation biases inherent in those models, while also providing a robust conversion prediction model. (Ya Zhang, Wei, and Ren 2014)

  • Bridge the gap between academia and the financial industry by using advanced machine learning techniques, such as deep neural networks, gradient-boosted trees, and random forests, to develop short-term statistical arbitrage strategies for the S&P 500 constituents. (I. J. Goodfellow, Warde-Farley, Mirza, et al. 2013)

  • Consider the online portfolio selection problem as a sequential decision problem, and study various state-of-the-art approaches like Follow-the-Winner’, ‘Follow-the-Loser’, ‘Pattern-Matching’ based methods, and ‘Meta-Learning Algorithms’.’ (Bin Li and Hoi 2012)

  • Carefully consider the complexities involved in planning and scheduling problems in the process industry, taking into account factors such as multi-purpose production units, sequence-dependent set-up times, non-preemptive processes, and multi-component flow and nonlinear blending when developing models and selecting appropriate solution techniques. (“Overview of Industrial Batch Process Scheduling” 2010)

  • Focus on developing algorithms that balance traffic and performance under smooth delivery constraints in order to effectively manage budgets in real-time bidding environments. (S. Agrawal, Wang, and Ye 2009)

  • Utilize composition formulas and integration-by-part rules for fractional integrals and Weyl derivatives to extend the classical Lebesgue-Stieltjes integral to a larger class of integrands and integrators of unbounded variation. (Zähle 1998)

  • Use multiple models and representations when analyzing complex datasets like those found in the telecom industry, and then combine these models using methods such as majority voting or Adaboost to improve overall prediction accuracy. (NA?)

  • Consider utilizing Support Vector Machines (SVMs) instead of Backpropagation (BP) for financial forecasting, as SVMs demonstrate superior performance across various evaluation metrics such as Normalized Mean Square Error (NMSE), Mean Absolute Error (MAE), Directional Symmetry (DS), Correct Up (CP) trend, and Correct Down (CD) trend. (NA?)

  • Carefully consider the implications of unbalanced data, non-stationarity, and appropriate evaluation metrics when developing and implementing credit card fraud detection algorithms. (NA?)

  • Focus on addressing the critical issues of class imbalance, non-stationarity, and model evaluation when developing effective credit card fraud detection systems using machine learning techniques. (NA?)

  • Utilise an adaptive strategy for maximising the click through rate (CTR) in online advertising, which relies on estimating the preference characteristic for a new request and proposing a relevant bid price based on the look-alike’ principle without employing any parametric models. (NA?)

  • Utilize machine learning methods to analyze large datasets in finance, specifically focusing on risk premium measurement, as these methods allow for greater flexibility in model specification and can lead to significant improvements in predictive accuracy. (NA?)

  • Consider implementing and reproducing previous studies as baselines before developing new deep learning models for stock market prediction, focusing on the latest advancements in the field within the past three years. (NA?)

Adversarial Machine Learning

  • Implement integrity verification measures to protect against split-view poisoning attacks and timing-based defenses to guard against frontrunning poisoning attacks in web-scale training datasets. (Carlini et al. 2023)

  • Carefully examine the effects of adversarial training on neural networks performance across various architectures, training methods, and datasets, paying close attention to the trade-offs between adversarial robustness and human-like behavior.’ (Gavrikov, Keuper, and Keuper 2023)

  • Carefully evaluate the potential for dual-use risks associated with instruction-following large language models (LLMs), as your enhanced capabilities can lead to significant economic incentives for malicious actors to exploit them for harmful purposes. (D. Kang et al. 2023)

  • Carefully consider the impact of small changes in prompt design on the output of generative models, as these can lead to significant differences in the resulting content. (Maus et al. 2023)

  • Conduct large-scale crowdsourced competitions to gather adversarial prompts against state-of-the-art LLMs, enabling a deeper understanding of your vulnerability to prompt hacking and informing the development of effective defense mechanisms. (Schulhoff et al. 2023)

  • Carefully consider the limitations and assumptions underlying your statistical analyses, and actively engage in sensitivity analyses to explore the robustness of your results to alternative scenarios. (Shavit 2023)

  • Use masked images as counterfactual samples to improve the robustness of fine-tuning models, specifically by masking either semantics-related or semantics-unrelated patches of the images based on class activation maps and then refilling the masked patches with patches from other images. (Yao Xiao et al. 2023)

  • Consider the potential security threats posed by backdoor attacks on prompt-based large language models and investigate ways to mitigate these risks. (Hongwei Yao, Lou, and Qin 2023)

  • Consider the potential vulnerabilities of deep vision models to adversarial attacks, particularly in security-sensitive applications, and explore ways to enhance your robustness. (Chenshuang Zhang, Zhang, Kang, et al. 2023)

  • Focus on creating robust, multi-prompt and multi-model attacks to effectively manipulate aligned language models, utilizing combined greedy and gradient-based discrete optimization techniques. (A. Zou et al. 2023)

  • Carefully consider the dual-use implications of generative AI (GenAI) technologies, such as large language models (LLMs) and diffusion models, as they possess significant potential for both beneficial and malicious applications. (Barrett et al. 2023)

  • Utilize an efficient approach for verifying specifications definable using Latent Variable Models (LVMs) in order to ensure the robustness of neural networks deployed in safety-critical applications. (Kouvaros et al. 2023)

  • Utilise machine learning models like BERT and XGBoost to analyse vast amounts of natural language data available on the internet to effectively identify and evaluate the level of cybersecurity threats and vulnerabilities in the healthcare sector. (Silvestri et al. 2023)

  • Consider extending MLOps to TinyMLOps when deploying machine learning models on edge devices, taking into account the unique challenges associated with edge deployment such as managing model versions, observability, pay-per-query business models, retraining and personalizing models, targeting a fragmented IoT landscape, and protecting the models intellectual property.’ (Leroux et al. 2022)

  • Consider implementing the MaxUp’ technique, which involves generating a set of augmented data with random perturbations or transforms, and minimizing the maximum, or worst case loss over the augmented data. This approach implicitly introduces a smoothness or robustness regularization against the random perturbations, thereby improving the generalization performance of machine learning models, particularly deep neural networks.’ (Guangyao Chen et al. 2021)

  • Utilise the proposed Dual Prior Alignment (DPA)’ network to automatically assess the naturalness of physical world attacks, thereby reducing human error and bias in the evaluation process. (Cherepanova et al. 2021)

  • Consider employing a cycle-consistent generative adversarial network (Cycle-GAN) for creating synthetic CAPTCHA data, which can help reduce the need for extensive manual labelling and enhance the efficiency of the overall attack process. (Chunhui Li et al. 2020)

  • Develop a novel approach called Image-based 2-dimensional Character Embedding Space (I2CES) to effectively defend against visual text attacks on neural networks, as it enables the model to capture more visual information and thus become more robust against such attacks. (Shengjun Liu, Jiang, and Wu 2020)

  • Pay attention to the potential issue of robust overfitting’, which refers to the phenomenon where continued training of adversarially robust deep learning models may lead to increased robust test losses, thus negatively impacting overall performance. (Rice, Wong, and Kolter 2020)

  • Consider implementing friendly adversarial training (FAT) in order to achieve adversarial robustness without sacrificing natural generalization. (Jingfeng Zhang et al. 2020)

  • Carefully consider the impact of structural attacks when developing graphical models like Associative Markov Networks (AMN) for classification tasks, and propose a novel bi-level program framework to optimize these models against such attacks. (K. Zhou and Vorobeychik 2020)

  • Consider various types of adversarial attacks and defences when evaluating the safety and reliability of deep neural networks, including poisoning and evasion attacks, white-box, black-box, and semi-white box attacks, and defence strategies such as gradient masking, robust optimization, and adversary detection. (Han Xu et al. 2020)

  • Consider using adversarial training to enhance the reliability of post-hoc explanation methods for graph neural networks (GNNs), as demonstrated by its effectiveness in improving representation extraction for GNNs and reducing the need for complex post-hoc explanation methods. (sivaraman et al. 2019)

  • Consider the potential impact of floating-point errors on the accuracy of gradient-based attacks when evaluating the robustness of deep learning models. (Carlini et al. 2019)

  • Consider developing a new optimization method for deriving backdoor triggers that directly minimizes individual pixel changes, rather than using a mask to define the set of pixels that ought to be perturbed. This approach can lead to triggers that require fewer input pixels to be perturbed, have a higher attack success rate, and are more robust, resulting in improved performance in both attack and defense scenarios. (A. Chan and Ong 2019)

  • Consider using parametric perturbations, particularly geometric transformations, to improve the efficiency and effectiveness of black-box adversarial attacks on video analysis systems. (J. Du et al. 2019)

  • Consider the potential for player domination’, where one player controls the outcome of a bargaining game, leading to non-convergence in certain situations. (D. Kang et al. 2019)

  • Utilise the Adaptive Diversity Promoting (ADP) regulariser to enhance the diversity within your ensemble model. This technique involves promoting the diversity amongst the non-maximal predictions of various members in the ensemble, thereby improving the overall robustness of the system. (T. Pang et al. 2019)

  • Consider the possibility of adversarial attacks on machine learning models, particularly those involving person detection, and develop strategies to mitigate your effects. (Thys, Ranst, and Goedemé 2019)

  • Carefully balance the trade-off between robustness and accuracy in your studies, taking into account factors like the natural error, boundary error, and the choice of surrogate loss function. (Hongyang Zhang et al. 2019)

  • Carefully consider the potential for adversaries to manipulate graph structures in order to evade detection systems, and develop appropriate countermeasures accordingly. (Binghui Wang and Gong 2019)

  • Use a combination of local and global perturbation budgets to ensure the robustness of graph convolutional networks against adversarial attacks. (Zügner and Günnemann 2019)

  • Utilise adversarial training methods to improve the robustness and accuracy of your machine learning models, particularly those involving gradient-boosted decision trees. (Calzavara, Lucchese, and Tolomei 2019)

  • Focus on understanding and applying various privacy-preserving techniques, including cryptography, differential privacy, and perturbation methods, to safeguard sensitive data during machine learning processes. (Al-Rubaie and Chang 2018)

  • Carefully consider the trade-off between reliability and stealthiness when developing adversarial attacks against text-to-image models, and propose a novel genetic-based optimization method to balance these competing objectives. (Alzantot et al. 2018)

  • Consider incorporating adversarial training (AT) as a regularization technique in your neural network models for improved robustness and accuracy in tasks like entity recognition and relation extraction. (Bekoulis et al. 2018)

  • Engage with technical research to investigate, prevent, and mitigate potential malicious uses of AI, taking into account ethical considerations and involving a wider range of stakeholders and domain experts. (Brundage et al. 2018)

  • Consider implementing the FreeLB algorithm for adversarial training in natural language processing tasks, as it demonstrates superior performance compared to traditional methods in terms of generalization and robustness. (P. Clark et al. 2018)

  • Consider the potential impact of adversarial attacks on graph neural networks, which can lead to misclassification and reduced performance, especially in areas such as finance and security. (H. Dai et al. 2018)

  • Participate in competitive environments to develop and test novel approaches for generating adversarial examples and defending against them, as this provides a more rigorous evaluation compared to traditional benchmarking methods. (Kurakin et al. 2018)

  • Develop a comprehensive intellectual property rights (IPR) protection framework for Generative Adversarial Networks (GANs) that includes both black-box and white-box settings, ensuring that the original GANs performance is preserved while being resistant to removal and ambiguity attacks. (Rouhani, Chen, and Koushanfar 2018)

  • Utilize the expressive capabilities of generative models like GANs to defend deep neural networks against adversarial attacks, without altering the classifier structure or training procedures, and without assuming knowledge of the process for generating adversarial examples. (Samangouei, Kabkab, and Chellappa 2018)

  • Develop novel algorithms for generating adversarial point clouds against deep neural networks for point cloud processing, focusing on both adversarial point perturbation and adversarial point generation techniques, while considering specific perturbation metrics tailored to the attacks in point clouds. (Xiang, Qi, and Li 2018)

  • Carefully consider the potential impact of poisoning attacks on feature selection algorithms, as these attacks can significantly compromise the efficacy of such algorithms, leading to poor generalization and potentially allowing malicious actors to evade detection. (Huang Xiao et al. 2018)

  • Utilise a convex outer approximation of the set of activations reachable through a norm-bounded perturbation, and develop a robust optimization procedure that minimises the worst case loss over this outer region, resulting in a deep network that is provably robust to any norm-bounded adversarial attack. (Athalye et al. 2017)

  • Carefully consider the sampling strategy employed during the transfer set construction phase, as it significantly impacts the effectiveness of the knockoff model in accurately emulating the target black box model. (Bhagoji, He, et al. 2017)

  • Consider adding a separate detector’ subnetwork to your deep neural networks, which is trained to distinguish between genuine data and data containing adversarial perturbations. This method shows promising results in detecting adversarial perturbations, even those that are quasi-imperceptible to humans, and can generalize to similar and weaker adversaries.’ (Metzen et al. 2017)

  • Utilize back-gradient optimization to efficiently compute gradients of interest through automatic differentiation, thereby reducing the complexity of poisoning attacks on deep learning algorithms. (Muñoz-González et al. 2017)

  • Adopt a principled adversarial training approach to improve the robustness of neural network models against adversarial attacks, utilizing a Lagrangian penalty formulation within a Wasserstein ball framework to augment model parameter updates with worst-case perturbations of training data. (A. Sinha et al. 2017)

  • Investigate the dimensionality of adversarial subspaces to understand the limits of transferability in machine learning models. (Tramèr, Papernot, et al. 2017)

  • Carefully balance the tradeoff between stealthiness and robustness in developing backdoor attacks against neural networks, as focusing solely on either aspect could lead to weaknesses that can be exploited by defenders. (Xinyun Chen et al. 2017)

  • Avoid assuming that combining multiple weak defences against adversarial examples will result in a strong defence, as the paper demonstrates that an adaptive adversary can still create adversarial examples with low distortion. (W. He et al. 2017)

  • Consider using adversarial training for visual saliency prediction, specifically by implementing a SalGAN model, which involves two competing networks: a generator that produces saliency maps from raw pixels of an input image, and a discriminator that determines if the produced saliency map is genuine or artificial. This approach enables the model to perform optimally across various saliency metrics. (Junting Pan et al. 2017)

  • Focus on developing ensemble adversarial training approaches to improve the robustness of machine learning models against adversarial attacks, particularly those involving black-box adversaries. (Tramèr, Kurakin, et al. 2017)

  • Investigate the robustness of multimodal neural networks against worst-case (i.e., adversarial) perturbations on a single modality, and propose an adversarially robust fusion strategy that trains the model to compare information coming from all the input sources, detect inconsistencies in the perturbed modality compared to the other modalities, and only allow information from the unperturbed modalities to pass through. (Arevalo et al. 2017)

  • Consider implementing a multi-adversarial domain adaptation (MADA) approach in order to capture multimode structures and achieve fine-grained alignment of different data distributions based on multiple domain discriminators, thereby improving the effectiveness of domain adaptation techniques. (Arjovsky, Chintala, and Bottou 2017)

  • Utilize adversarial attacks as a function evaluation tool to search for neural architectures that can resist such attacks automatically, thereby improving the overall robustness of neural networks. (Brock et al. 2017)

  • Use generative models to represent the low-dimensional data manifold, allowing for optimization of adversarial patches within this manifold, thereby reducing the gap between the responses of substitute models and target models and enhancing the transferability of adversarial patches. (T. B. Brown et al. 2017)

  • Use a combination of theoretical guarantees and empirical illustrations to demonstrate the robustness of your models against adversarial attacks, particularly focusing on the use of smoothed classifiers and tight certificates of adversarial robustness. (Carlini et al. 2017)

  • Apply image transformations such as bit-depth reduction, JPEG compression, total variance minimization, and image quilting before feeding images to a convolutional network classifier to effectively defend against adversarial-example attacks on image-classification systems. (C. Guo et al. 2017)

  • Adopt a robust optimization approach to studying adversarial robustness in neural networks, allowing for a principled and unifying view on previous work and providing a concrete security guarantee against any adversary. (Madry et al. 2017)

  • Focus on developing a unified framework for detecting out-of-distribution samples and adversarial attacks, leveraging the power of generative classifiers and the Mahalanobis distance-based score to achieve state-of-the-art performance in both cases. (Amodei et al. 2016)

  • Carefully evaluate the robustness of neural networks by developing stronger attacks and testing your effectiveness on defended models, rather than solely relying on existing defenses. (Carlini and Wagner 2016)

  • Focus on developing white-box membership inference attacks specifically tailored to exploit the privacy vulnerabilities of the Stochastic Gradient Descent (SGD) algorithm, rather than simply extending black-box attacks to the white-box setting. (Konečný et al. 2016)

  • Use the standardized reference implementation provided by the cleverhans library to ensure accurate and comparable benchmarks of your machine learning models performance in the adversarial setting.’ (Papernot, Faghri, et al. 2016)

  • Focus on developing and testing adversarial input sequences for recurrent neural networks, as these models are particularly susceptible to adversarial manipulations due to your ability to handle sequential data and the presence of cyclical computations in your architectures. (Papernot, McDaniel, et al. 2016)

  • Focus on generating a diverse set of adversarial images, rather than solely focusing on the closest adversarial images, to enhance the training set and improve the accuracy and robustness of learning models. (Rozsa, Rudd, and Boult 2016)

  • Consider the potential for membership inference attacks when developing machine learning models, particularly when working with sensitive data, and explore ways to mitigate this risk through various strategies such as limiting the models predictions to top k classes, decreasing the precision of the prediction vector, increasing its entropy, or using regularization while training the model.’ (Shokri et al. 2016)

  • Carefully consider the potential vulnerabilities of machine learning models deployed through publicly accessible query interfaces, particularly those that return rich outputs like confidence values, and develop robust countermeasures to prevent model extraction attacks. (Tramèr et al. 2016)

  • Utilise a combination of advanced techniques such as deep learning, adversarial objectives, and mixed-effects modelling to identify and control for potential confounding factors in your analysis, thereby improving the accuracy and reliability of your findings. (Abadi et al. 2016)

  • Consider the potential impact of multiple source domains and category shifts when performing unsupervised domain adaptation, and they can address these issues through the use of a deep cocktail network (DCTN) approach that combines multi-way adversarial learning and integration of source-specific perplexity scores. (Bousmalis et al. 2016)

  • Utilise a combined approach of generative image space alignment and latent representation space alignment for successful domain adaptation, particularly in cases involving significant visual domain shifts. (Dumoulin et al. 2016)

  • Consider using the Private Aggregation of Teacher Ensembles (PATE) methodology to ensure strong privacy guarantees for sensitive training data in machine learning applications. (Papernot, Abadi, et al. 2016)

  • Consider the possibility of adversaries exploiting the efficiency vulnerability in dynamic neural networks (DyNNs) through the injection of universal efficiency backdoors, leading to a false sense of efficiency and potential misuse of computational resources. (Besse et al. 2015)

  • Focus on understanding the relationship between the smoothness and dimensionality of generative models and the robustness of classifiers against adversarial perturbations, as these factors significantly impact the effectiveness of classifiers in handling small, imperceptible alterations in data. (K. He et al. 2015a)

  • Utilise “defensive distillation” as a defence mechanism against adversarial samples in deep neural networks (DNNs), as it significantly reduces the effectiveness of adversarial samples and increases the minimum number of features required to be modified to create adversarial samples. (Papernot et al. 2015)

  • Use a distribution quantile bound for activation values and a polynomial barrier loss function to effectively create adversarial attacks on deep content features, achieving a state-of-the-art trade-off between attack success rate and imperceptibility. (I. J. Goodfellow, Shlens, and Szegedy 2014)

  • Focus on developing a hybrid approach combining gradient-based white-box methods and zeroth-order optimization in black-box methods to effectively address the model mismatch issue in transfer-based black-box attacks. (I. J. Goodfellow, Shlens, and Szegedy 2014)

  • Utilize meta-gradients to address the bilevel optimization problem inherent in poisoning attacks against graph neural networks, effectively treating the graph structure as a hyperparameter to optimize. (F. Agostinelli et al. 2014)

  • Investigate the availability of LiDAR detection pipelines, specifically focusing on how adversarial perturbations affect latency rather than just integrity, and developing techniques like SlowLiDAR to maximise detection runtime. (S. Gu and Rigazio 2014)

  • Consider using DP-FTRL, a differentially private variant of the Follow-the-Regularized-Leader (FTRL) algorithm, for online convex optimization (OCO) tasks, as it offers improved regret guarantees and flexibility in data access patterns compared to DP-SGD. (Duchi, Jordan, and Wainwright 2013)

  • Focus on developing a robust classifier for deep neural networks by defining a Max-Mahalanobis distribution (MMD) and proposing a novel Max-Mahalanobis linear discriminant analysis (MM-LDA) network, which explicitly maps a complicated data distribution in the input space to a MMD in the latent feature space and then applies LDA to make predictions. (Diederik P. Kingma and Welling 2013)

  • Consider utilising a generative adversarial network (GAN) to efficiently generate and defend against adversarial examples in deep neural networks (DNNs), thereby increasing your adversarial stability. (Szegedy et al. 2013)

  • Consider incorporating robustness against adversarial training data as a critical aspect in the development of learning algorithms, as demonstrated by the successful implementation of a poisoning attack against Support Vector Machines (SVM) using a gradient ascent strategy. (Biggio, Nelson, and Laskov 2012)

  • Be aware of potential vulnerabilities introduced by using statistical machine learning for decision-making in large-scale systems, particularly regarding the risk of adversarial manipulation of training data. (Barreno et al. 2006)

  • Carefully consider potential adversarial attacks on statistical machine learning techniques, particularly in the context of network security, and develop robust defenses against such attacks. (NA?)

  • Adopt a comprehensive approach to evaluating secure learning systems, involving identifying various classes of attacks, assessing the resilience of existing systems against these attacks, and exploring potential defenses against them. (NA?)

  • Consider using a large feature set and advanced machine learning techniques to accurately differentiate regular documents from deceptive documents, achieving up to 96.6% accuracy (F-measure) in doing so. (NA?)

  • Proactively design crowdsourcing tasks to be inherently resistant to cheating, rather than relying on post-task detection and rejection of cheaters. (NA?)

  • Combine homomorphic encryption and Yao garbled circuits to achieve optimal performance in privacy-preserving ridge regression, as demonstrated by the authors successful implementation and testing on real datasets.’ (NA?)

  • Proactively protect your machine learning models by identifying potential vulnerabilities before they are exploited, investigating the impact of corresponding attacks, and devising appropriate countermeasures if needed. (NA?)

  • Utilize a leveled homomorphic encryption scheme to ensure the confidentiality of training and test data during delegation of machine learning algorithm execution to a computing service. (NA?)

  • Carefully consider the potential for model inversion attacks when developing machine learning algorithms, especially in privacy-sensitive applications, and implement appropriate countermeasures to protect sensitive information. (NA?)

  • Utilize the moments accountant methodology for privacy accounting in differentially private stochastic gradient descent (SGD) algorithms, as it offers a tighter bound compared to the strong composition theorem, thus providing improved privacy protection while maintaining model accuracy. (NA?)

  • Consider developing methods to create physically realizable and inconspicuous attacks on facial biometric systems, focusing on dodging and impersonation tactics, in order to better understand and mitigate potential threats to these technologies. (NA?)

  • Utilize the proposed protocols for privacy-preserving machine learning, specifically for linear regression, logistic regression, and neural network training, within the two-server model, to achieve scalability and efficiency improvements. (NA?)

  • Consider the potential for membership inference attacks when developing machine learning models, particularly when working with sensitive data, and explore ways to mitigate this risk through techniques such as limiting the models predictions to top k classes, decreasing the precision of the prediction vector, increasing its entropy, or using regularization while training the model.’ (NA?)

  • Focus on developing robust, reliable, and transparent machine learning systems capable of withstanding various types of hazards, including adversarial attacks, long-tailed events, and hidden functionalities, while ensuring that these systems are aligned with human values and goals. (NA?)

  • Be aware of the potential for obfuscated gradients in your defenses against adversarial examples, and take steps to ensure that your defenses are truly robust rather than just appearing to be so due to gradient masking. (NA?)

  • Utilise IoT-specific network behaviours when selecting features for machine learning algorithms to accurately detect DDoS attacks in IoT network traffic. (NA?)

  • Consider the potential for adversarial attacks on machine learning algorithms, particularly in areas such as computer security, and develop appropriate countermeasures to mitigate these risks. (NA?)

  • Consider leveraging the sensitivity of modern machine learning algorithms to input perturbations in order to design “robust objects,” i.e., objects that are explicitly optimized to be confidently detected or classified, thereby significantly improving vision models performance and robustness.’ (NA?)

  • Focus on developing robust defense mechanisms against adversarial attacks in artificial intelligence systems, particularly through modifying data, altering models, and employing auxiliary tools. (NA?)

  • Focus on understanding and addressing the presence of non-robust features in your datasets, as these features contribute to the creation of adversarial examples and impact the robustness of machine learning models. (NA?)

  • Explore the potential of differential evolution (DE) as a powerful tool for generating adversarial examples in deep neural networks, particularly in limited scenarios where only one pixel can be modified, as demonstrated by the impressive performance of the proposed method in fooling multiple types of networks. (NA?)

  • Consider the potential impact of label noise and improper representation learning on adversarial vulnerability in deep neural networks, as these factors can significantly affect the performance and robustness of the models. (NA?)

  • Focus on improving the performance of MaxCosine and MaxNorm through modifications to standard training, such as incorporating a cosine classifier and adjusting training losses, to effectively utilize the Decoupling MaxLogit (DML) method for out-of-distribution detection. (NA?)

  • Carefully consider the potential vulnerabilities of neural ranking models (NRMs) to adversarial attacks, particularly in the context of decision-based black-box attack settings, and develop appropriate defense mechanisms accordingly. (NA?)

  • Utilise the objective perturbation’ method rather than the previously established ‘output perturbation’ method when developing privacy-preserving machine learning algorithms. (NA?)

  • Consider the potential for clean-label backdoor attacks in prompt-based learning systems, as demonstrated by the ProAttack method, which uses the prompt itself as a trigger to manipulate the output of downstream tasks. (NA?)

  • Carefully consider the trade-offs between proactive and reactive security solutions when applying data mining and machine learning techniques to cybersecurity problems. (NA?)

Defense Mechanisms Against Adversaries

  • Consider implementing the Signed-Prompt’ method as a defense strategy against prompt injection attacks in large language model integrated applications, whereby sensitive instructions are signed by authorized users, allowing the model to discriminate between trusted and untrusted instruction sources.’ (Suo 2024)

  • Consider incorporating clean graphs from similar domains as a means of building robust graph neural networks (GNNs) capable of mitigating the negative effects of adversarial attacks. (X. Tang, Li, et al. 2020)

  • Focus on improving the generalization of adversarial training through domain adaptation, specifically by treating adversarial training on Fast Gradient Sign Method (FGSM) as a domain adaptation task with limited number of target domain samples, thereby enabling the adversarially trained model to perform well on adversarial examples crafted by FGSM and showing great generalization on other adversaries. (C. Song et al. 2018)

  • Utilize a collaborative multi-task training framework to improve the robustness of your deep neural networks against adversarial attacks. (Derek Wang et al. 2018)

  • Utilise a novel training procedure called Reverse Cross Entropy (RCE) to enhance the ability of deep learning models to differentiate between adversarial and normal examples. (Bhagoji, Cullina, et al. 2017)

Fairness And Bias Mitigation

  • Employ a calibrated projection matrix to debias vision-language foundation models, enabling improved group robustness in both discriminative and generative models without requiring additional data or training. (C.-Y. Chuang et al. 2023)

  • Consider integrating equal opportunity directly into your training objectives to effectively reduce bias while maintaining high performance in classification tasks. (A. Shen et al. 2022)

  • Develop a general framework for Fair Survival Time Prediction (FAST) that directly achieves Demographic Parity (DP) between the predicted survival time and sensitive attributes, instead of DP on model outputs, to ensure fairness in survival analysis models. (I. Y. Chen et al. 2021)

  • Focus on developing fairness-aware outlier detection models that balance between accuracy and fairness, taking into account multiple desiderata such as detection effectiveness, treatment parity, statistical parity, group fidelity, and base rate preservation. (Shekhar, Shah, and Akoglu 2021)

  • Consider implementing a flexible and scalable framework like LiFT to effectively measure and mitigate bias in large-scale AI applications throughout the entire ML lifecycle, including before, during, and after training, as well as during online serving. (Vasudevan and Kenthapadi 2020)

  • Use the proposed Data Shapley’ approach to determine the value of each data point in a dataset, based on its contribution to the overall performance of a machine learning model. (A. Ghorbani and Zou 2019)

  • Aim to develop algorithms that generate predictions that are statistically independent of protected attributes like race, ensuring fairness in the context of recidivism prediction. (Johndrow and Lum 2019)

  • Focus on developing an extensible, open-source toolkit like AI Fairness 360 (AIF360) that brings together a comprehensive set of bias metrics, bias mitigation algorithms, bias metric explanations, and industrial usability, enabling stronger collaboration between AI fairness researchers and practitioners across various industries. (Bellamy et al. 2018)

  • Aim to minimize the discrepancy between the factual and counterfactual distributions in your study designs, thereby reducing bias and improving the validity of your findings. (F. D. Johansson, Shalit, and Sontag 2016)

  • Consider using distributionally stable algorithms to mitigate the impact of sample selection bias on the accuracy of your models. (Cortes et al. 2008)

  • Consider the three main components of concept drift - detection, understanding, and adaptation - when developing methodologies and techniques for handling concept drift in machine learning. (NA?)

  • Develop and utilize model cards’, which are comprehensive summaries of machine learning models that include information on the model’s type, intended use, factors influencing performance, evaluation data, and ethical considerations, thereby enabling stakeholders to make informed decisions about model selection and deployment.’ (NA?)

  • Carefully consider and account for sex and gender differences in your biomedical AI technologies to minimize biases and maximize precision medicines potential for improving health outcomes.’ (NA?)

Privacy Preservation

  • Consider implementing a privacy-preserving machine learning as a service system like Chiron, which utilizes SGX enclaves and Ryoan sandboxes to maintain data privacy and model confidentiality while offering black-box access to trained models. (Hunt et al. 2018)

  • Carefully choose the prior distribution in order to ensure both robustness and privacy in Bayesian inference, as this can lead to significant improvements in the overall performance of the analysis. (Dimitrakakis et al. 2013)

Probability And Statistical Learning Theory

  • Utilise the Monte Carlo method for drawing parameter values from a distribution defined on the structural parameter space of an equation system. This method allows for a flexible choice of prior distributions and doesnt exponentially increase the number of elementary operations with the increasing number of parameters. Furthermore, it helps overcome some challenges associated with applying Bayesian methods to mid-size models.’ (Minghao Li et al. 2023)

  • Utilize PAC-Bayes bounds to analyze the generalization performance of your models, particularly when dealing with complex predictive systems such as neural networks. (Alquier 2021)

  • Utilise negative controls - observed covariates with specific relationships to the action and outcome - to address unmeasured confounding in causal inference, particularly when traditional approaches like controlling for all confounders are impractical or impossible. (Kallus, Mao, and Uehara 2021)

  • Utilise a low-rank semiparametric Bayesian spatial mixed-effects linear model to effectively manage large, highly nonstationary spatiotemporal datasets. (Hazra and Huser 2021)

  • Carefully consider the timing of imputation in your cross-validation process, as performing unsupervised imputation before cross-validation (I→CV) can potentially lead to biased estimation of a modeling pipeline’s generalization error and negatively affect the selection of tuning parameters. (Jaeger, Tierney, and Simon 2020)

  • Consider using Empirical Bayesian Kriging (EBK) as a reliable automatic interpolator for spatial data, especially when dealing with complex data sets, as it accounts for estimation error in the semivariogram model and offers improved accuracy compared to traditional kriging methods. (Krivoruchko and Gribov 2019)

  • Carefully evaluate multiple models and compare your performance using objective metrics like leave-one-out cross-validation to identify the best-fit model for explaining human confidence behavior. (Adler and Denison 2018)

  • Use PAC-Bayesian risk bounds to optimize your models, which involves balancing the empirical expected loss against the Kullback-Leibler divergence. (P. Germain et al. 2016)

  • Utilize the Tracy-Widom law of order 1 to estimate the distribution of the largest eigenvalue in principal components analysis, as it offers accurate predictions even for small values of n and p. (Hürlimann 2015)

  • Utilize anti-concentration inequalities for maxima of Gaussian random vectors to establish bounds on the Kolmogorov distance between maxima of Gaussian random vectors, which is crucial in various areas including mathematical statistics and high-dimensional statistical inference. (Chernozhukov, Chetverikov, and Kato 2014)

  • Utilize the anisotropic local law when analyzing random matrices, as it provides a more accurate representation than traditional isotropic local laws. (Knowles and Yin 2014)

  • Utilize the Linear Noise Approximation (LNA) when studying complex chemical networks, as it allows for more accurate predictions and analysis while being computationally efficient. (Finkenstädt et al. 2013)

  • Consider combining both GPS data and trip start and end locations and times to improve the accuracy of ambulance travel time predictions, while accounting for potential biases caused by the GPS sampling scheme. (Westgate et al. 2013)

  • Incorporate heterogeneous operating conditions into your analyses of accelerated life testing (ALT) data to better predict field failures and optimize ALT experiment designs. (Z.-S. Ye, Hong, and Xie 2013)

  • Carefully choose and justify your preferred method for measuring predictive accuracy, recognizing that no perfect solution exists and that different methods have varying levels of bias and computational complexity. (Andrew Gelman, Hwang, and Vehtari 2013)

  • Utilize the cavity method, borrowing insights from the study of mean field spin glasses, to analyze the extremal process of branching Brownian motion. (Arguin, Bovier, and Kistler 2012)

  • Utilise the Gumbel process, a stochastic process recently introduced in mathematical statistics, to convert the problem of sampling from a continuous distribution into an optimization problem over continuous space. (Dymetman, Bouchard, and Carter 2012)

  • Utilise max-stable processes for modelling spatial dependence in extreme events, particularly in situations where traditional methods like latent processes or copulas fail to adequately capture the complexity of the phenomenon being studied. (Blanchet and Davison 2011)

  • Consider extending your analysis beyond traditional stationary models by incorporating nonstationary nested SPDE models, which offer greater flexibility and improved computational efficiency while still retaining desirable properties like easy nonstationary extensions and applicability to data on general smooth manifolds. (Bolin and Lindgren 2011)

  • Focus on developing a “conditioning-free” process, called the effective or driven process, which shares the same typical states as the conditioned process in the stationary limit, allowing for better understanding and analysis of complex systems. (Chetrite and Gupta 2011)

  • Utilize a population Monte Carlo correction to improve the accuracy of your statistical models, particularly when dealing with complex scenarios like population genetics. (Beaumont et al. 2009)

  • Consider using the Maximum Mean Discrep Question (MMD) as a test statistic for determining if two samples come from different distributions, particularly when working with high dimensional data and limited sample sizes. (Gretton et al. 2008)

  • Optimize the performance of your Markov Chain Monte Carlo (MCMC) algorithms by employing controlled MCMC techniques, which involves adjusting the parameters of the algorithm based on the observed data during the sampling process, ultimately improving the accuracy and efficiency of the estimation. (Andrieu and Thoms 2008)

  • Consider using the proposed Widely Applicable Bayesian Information Criterion (WBIC) instead of traditional methods like BIC, especially when dealing with singular statistical models, as it provides a more accurate estimation of the Bayes Free Energy even when the true distribution is unknown. (Saito 2007)

  • Focus on developing a comparison theorem for Backward Stochastic Differential Equations (BSDEs) with unbounded terminal conditions under the assumption of convexity of the generator with respect to the variable z. (Briand and Hu 2007)

  • Focus on developing optional decompositions that hold simultaneously for all equivalent martingale measures, allowing them to analyze hedging problems with constrained portfolios effectively. (Föllmer and Kramkov 2006)

  • Utilize a localization procedure combined with a priori bounds to establish the existence of solutions to BSDEs with quadratic growth and unbounded terminal value. (Briand and Hu 2006)

  • Utilize the propagation-separation (PS) approach for local likelihood estimation across various nonparametric models, allowing for flexible and adaptive local neighborhoods around each design point, ultimately improving the accuracy and efficiency of your estimates. (Polzehl and Spokoiny 2005)

  • Carefully examine the boundary behavior of censored stable processes in non-smooth open sets, particularly regarding whether the process approaches the boundary in a finite time and how harmonic functions corresponding to the censored process behave near the boundary. (Bogdan, Burdzy, and Chen 2003)

  • Utilize the product of p-values as a test statistic for combining evidence from multiple independent sources, such as motif scores in sequence homology searches, to increase the accuracy and sensitivity of your analyses. (Bayat 2002)

  • Use the concept of filtration-consistency when defining nonlinear expectations, as it ensures that these expectations preserve monotonicity and constants, and allows for the derivation of conditional nonlinear expectations and nonlinear martingales. (Coquet et al. 2002)

  • Use the Robinson-Schensted-Knuth (RSK) correspondence to establish a bijection between matrices and pairs of semistandard Young tableaux, allowing for more efficient analysis of complex data structures. (Baryshnikov 2001)

  • Adopt a Bayesian statistical approach to photometric redshift estimation, as it enables the incorporation of prior knowledge, improves the accuracy of redshift estimation, and provides a robust measure of uncertainty. (Benitez 2000)

  • Utilise a simple Monte Carlo approach to approximate Bayesian credible and Highest Probability Density (HPD) intervals when dealing with complex models involving analytically intractable integrals. (M.-H. Chen and Shao 1999)

  • Utilize a purely probabilistic approach to study forward-backward stochastic differential equations and your connection with quasilinear parabolic partial differential equations, allowing the forward equation to be degenerate and under certain natural monotonicity conditions. (Pardoux and Tang 1999)

  • Utilize the Monotonic Limit Theorem of BSDE and the nonlinear decomposition theorem of Doob-Meyers type to analyze sequences of RCLL supersolutions of a backward stochastic differential equations (BSDE) and determine whether they converge monotonically up to a certain process, thereby providing insights into various fields including finance and economics.’ (Shige Peng 1999)

  • Consider applying large deviation principles to analyze the spectral measures of random matrices, particularly those governed by Wigners semicircular law, as these methods provide valuable insights into the behavior and properties of these complex systems.’ (Arous and Guionnet 1997)

  • Focus on analyzing the relationship between the evidence and generalization in the Bayesian setting, specifically examining the impact of the Occam factor on generalization performance. (Shawe-Taylor and Williamson 1997)

  • Utilize the Euler discretization scheme with step T/n for the approximation of stochastic differential equations, while considering the density of the law of X^n_T and comparing it to the density of the law of X_T, to ensure accurate modeling and prediction. (Bally and Talay 1996)

  • Focus on using the optional decomposition method for analyzing positive supermartingales within the context of a specific family of measures, rather than attempting to find a universal decomposition applicable across different families of measures. (Kramkov 1996)

  • Focus on deriving the existence of a smooth density for Yt in a framework that allows for countably supported measures μ, using a duality formula to estimate the characteristic function of F, and applying this to the specific case of F = Yt. (Picard 1996)

  • Consider using nonparametric estimators of autocovariance for stationary random fields, especially those that possess the property of being themselves autocovariances, as they enable the construction of bootstrap confidence intervals for unknown parameters and do not require assumptions such as isotropy or monotonicity. (Hall and Patil 1994)

  • Utilize a novel class of backward stochastic differential equations called doubly stochastic’, which can effectively represent the solution of a wide range of systems of quasi-linear parabolic SPDEs, thereby providing a powerful tool for studying these complex mathematical models.’ (Pardoux and Peng 1994)

  • Consider defining forward, backward, and symmetric stochastic integrals using a limit procedure, which extends Ito, backward, and Stratonovich integrals, respectively, while explicitly highlighting your forward’ nature and allowing for non-causal stochastic integration with respect to more general integrators than just Brownian motion. (Russo and Vallois 1993)

  • Carefully consider the choice of your statistical model when analyzing data, taking into account factors such as sample size, measurement error, and potential confounding variables. (Albeverio and Röckner 1991)

  • Adapt your maximum likelihood estimation method for imperfectly observed Gibbsian fields on a finite lattice using a novel algorithm based on the previous work of Younes [28 (Younes 1989)

  • Utilize the Feynman-Kac formula, a classical technique since McKean, to solve the KPP partial differential equation and study the large deviations of the Brownian Branching Motion model. (Chauvin and Rouault 1988)

  • Consider using Skorohods integral when dealing with stochastic processes, as it provides a more flexible framework compared to traditional methods while still maintaining mathematical rigor.’ (Nualart and Pardoux 1988)

  • Focus on removing unnecessary assumptions or conditions, like the admissibility condition in the study of Skorohod equations, to improve the robustness and applicability of your models. (Saisho 1987)

  • Consider the possibility of cube root asymptotics in your statistical models, particularly when dealing with sharp-edged effects, and use appropriate methods to handle them. (Kliemann 1987)

  • Use geometric techniques to analyze the convex hull of the likelihood set and its support hyperplanes to understand the existence, support size, likelihood equations, and uniqueness of the maximum likelihood estimator of a mixing distribution. (Kliemann 1987)

  • Use geometric techniques to analyze the convex hull of the likelihood set and its support hyperplanes to understand the existence, support size, likelihood equations, and uniqueness of the maximum likelihood estimator of a mixing distribution. (NA?)

  • Consider the possibility of cube root asymptotics in your statistical models, particularly when dealing with sharp-edged effects, and use appropriate methods to handle them. (NA?)

  • Prioritize stability over uniform convergence when considering learnability in the General Learning Setting, as it is a more powerful concept for characterizing learnability. (NA?)

  • Utilize the PAC-Bayesian theorem to establish distribution-free generalization error bounds for approximate Bayesian Gaussian process classification techniques, thereby providing a strong learning-theoretical justification for your use. (NA?)

  • Carefully evaluate the sensitivity of your chosen Bayesian synthetic likelihood (BSL) method to its tuning parameter n, the multivariate normal assumption, and computational efficiency, especially when comparing it to alternatives like approximate Bayesian computation (ABC). (NA?)

  • Prioritize the use of AUC (Area Under Curve) over accuracy in evaluating learning algorithms because AUC is a statistically consistent and more discriminating measure than accuracy, leading to improved ranking and ultimately greater net profit in practical applications. (NA?)

  • Utilise a Bayesian approach to integrate out the regularisation parameter in sparse logistic regression, thereby significantly increasing the efficiency of the algorithm without compromising its effectiveness. (NA?)

  • Focus on developing ranking rules based on U-processes, which offer superior performance compared to other approaches, particularly in situations where low-noise conditions exist. (NA?)

  • Carefully choose and evaluate the appropriateness of various imputation methods for handling missing data in your datasets, considering factors like the proportion of missing data, the type of variables involved, and the specific context of the study. (NA?)

  • Focus on developing data-dependent upper confidence bounds on the excess risk of empirical risk minimizers, which rely solely on the observed sample and do not require explicit knowledge of the underlying distribution. (NA?)

  • Consider both deterministic and statistical complexities as complementary approaches to understanding the behavior of physical systems, recognizing that while deterministic complexity measures degrees of randomness, statistical complexity measures degrees of structural organization. (NA?)

  • Utilize Approximate Bayesian Computation (ABC) to estimate the posterior distribution of parameters in complex simulation models without explicit likelihood functions, thus enabling accurate statistical inferences even in the absence of traditional analytical solutions. (NA?)

  • Consider using the Fast Unconstrained Bayesian AppRoximation (FUBAR) method for analyzing large datasets involving natural selection, as it provides a fast and accurate alternative to existing methods, reducing the risk of model misspecification and improving the identification of sites experiencing positive and purifying selection. (NA?)

  • Use caution when applying modern modelling techniques like SVM, NN, and RF in medical prediction problems, as these methods require significantly more events per variable to achieve a stable AUC-value compared to classical techniques like LR and CART, and thus should only be considered when very large datasets with many events are available. (NA?)

  • Use distance-induced kernels to resolve the issue of nonintegrability of weight functions in order to establish the link between RKHS-based dependence measures and the distance covariance. (NA?)

  • Understand the distinction between statistical inference and machine learning, and utilise appropriate techniques depending on whether your primary aim is to create a mathematical model of the data generation process for understanding or hypothesis testing (inference), or to forecast unobserved outcomes or future behaviour (prediction). (NA?)

  • Focus on understanding the difference between the error rate of a classification function and the area under the ROC curve (AUC) of a ranking function, as they require separate analyses and are not interchangeable indicators of model performance. (NA?)

  • Utilize a novel algorithmic approach to estimate high-dimensional inverse covariance matrices by leveraging the connection between multivariate linear regression and entries of the inverse covariance matrix, while taking advantage of the sparsity of the problem. (NA?)

  • Differentiate between aleatoric and epistemic uncertainty in machine learning, recognising that aleatoric uncertainty arises from inherent randomness in the data generating process, whilst epistemic uncertainty results from a lack of knowledge about the best model. (NA?)

  • Choose selection rules that lead to logically consistent methods of inference, which can be achieved through satisfying certain natural postulates. (NA?)

Bayesian Inference

  • Be cautious when relying solely on qualitative signatures derived from the Bayesian model of confidence, as these signatures often depend on hidden assumptions and are not necessarily unique to the Bayesian model. (NA?)

Information Theory

  • Use an information-theoretic approach to learning a Mahalanobis distance function by minimizing the differential relative entropy between two multivariate Gaussians under constraints on the distance function, which leads to a Bregman optimization problem that can be solved efficiently without requiring eigenvalue computations or semi-definite programming. (NA?)

Entropy And Mutual Information

  • Carefully consider the type of relationship being examined when choosing a feature selection method, as different methods perform differently depending on whether the relationship is linear or non-linear, and whether it involves multiple variables. (NA?)

Knowledge Representation And Reasoning

  • Adopt a unifying framework for supervised descriptive rule discovery, encompassing contrast set mining, emerging pattern mining, and subgroup discovery, to optimize rule coverage and precision in your analyses. (Xiaoyu Wang and Benning 2023)

  • Focus on developing methods for accurately extracting knowledge graphs from language models, ensuring high precision and recall rates, while considering factors like entity and relation paraphrasing, and utilizing techniques like few-shot in-context learning. (R. Cohen et al. 2023)

  • Employ a combination of techniques including Transformers, holistic reasoning, and soft label editing to effectively align entities within and across Knowledge Graphs (KGs) while taking into account various contextual factors such as relation, path, and neighborhood. (Xin et al. 2022)

  • Prioritize the development of computational causal inference (CompCI) software that is scalable, performant, and robust, enabling the integration of causal inference into large engineering systems and improving overall research agility. (J. C. Wong 2020)

  • Aim to create counterfactual visual explanations for deep computer vision systems, which involve identifying how regions of an input image would need to change in order for the system to produce a specified output, thus enabling greater interpretability and discrimination. (Y. Goyal et al. 2019)

  • Utilise the “Generalised Robinson Decomposition” (GRD) methodology when dealing with structured treatments like graphs, images, or texts. This method offers three significant benefits: firstly, it isolates the causal estimand, reducing regularisation bias; secondly, it permits the insertion of any supervised learning model for learning purposes; thirdly, it provides a quasi-oracle convergence guarantee under mild assumptions. The authors demonstrated the superiority of your approach (Athey, Tibshirani, and Wager 2019)

  • Carefully evaluate and compare counterfactual explanation generation algorithms based on multiple properties, including model access level, computational efficiency, interpretability, and ability to handle missing data, in order to select the most suitable algorithm for your specific application. (Adadi and Berrada 2018)

  • Model the problem of noun phrase (NP) and relation phrase canonicalization jointly rather than sequentially, utilizing relevant side information in a principled manner. (Vashishth, Jain, and Talukdar 2018)

  • Utilize an iterative approach when incorporating logic rules into knowledge graph embedding, allowing for improved transfer of knowledge from rules to embeddings. (S. Guo et al. 2017)

  • Utilise a novel scheme for both interpretation and explanation in deep neural networks, allowing for automated identification of internal features relevant for the set of classes considered by the model, without reliance on additional annotations. (Karpathy and Fei-Fei 2017)

  • Consider using a hybrid human-machine framework when dealing with large-scale knowledge base integration, specifically focusing on entity alignment, to improve both quality and cost-effectiveness. (Y. Zhuang et al. 2017)

  • Utilise multiple signals to identify areas of completeness in knowledge bases, and subsequently employ a rule mining approach to predict where facts might be missing. (Galárraga et al. 2017)

  • Utilise BetaE, a probabilistic embedding framework, when attempting to answer arbitrary first-order logic (FOL) queries over knowledge graphs (KGs). (Xiang Li, Vilnis, and McCallum 2017)

  • Utilize neural link predictors to identify missing edges in large scale Knowledge Graphs, and develop frameworks for efficiently answering complex queries on incomplete Knowledge Graphs by translating each query into an end-to-end differentiable objective, where the truth value of each atom is computed by a pre-trained neural link predictor. (Himmelstein et al. 2016)

  • Employ a compositional training’ objective when working with knowledge graphs embedded in vector spaces, as it significantly enhances the models’ capacity to accurately respond to path queries and provides a novel form of structural regularization, thereby enhancing performance across all base models.’ (Guu, Miller, and Liang 2015)

  • Carefully consider the impact of dataset shift on your models, particularly in relation to causality and conditional vs. unconditional models, and explore strategies like importance reweighting and local modelling to mitigate its effects. (S. Amos 2008)

  • Carefully consider the structure of your data and the relationships between variables when selecting appropriate methods for analysis. (“Knowledge Discovery in Databases: PKDD 2003” 2003)

  • Consider using hierarchical knowledge representation to facilitate decision-making and system management, particularly in complex systems where multiple levels of abstraction can aid in problem solving and efficient resource allocation. (Rasmussen 1985)

  • Adopt a subjective Bayesian inference method for rule-based inference systems, which combines the benefits of formal and informal approaches while accounting for inconsistencies inherent in collections of subjective statements. (Duda, Hart, and Nilsson 1976)

  • Consider applying the principles of algorithmic information theory to better understand and measure the power of formal axiomatic systems, thereby gaining insights into the limitations and potential improvements of various mathematical models. (Shelah 1974)

  • Focus on developing a comprehensive understanding of your domain, including intensional and extensional semantics, before attempting to create effective knowledge representation and reasoning systems. (“Experiments with the Graph Traverser Program” 1966)

  • Utilise the Structure-Mapping Engine (SME) toolkit to simplify experimentation and improve efficiency in cognitive simulation studies and machine learning systems, thereby enhancing the quality of your research outputs. (NA?)

  • Consider utilizing a Terminology Server to effectively integrate and standardize medical language and information systems, enabling seamless communication among various clinical applications while reducing complexity and improving efficiency. (NA?)

  • Carefully consider the descriptive, rhetorical, inferential, and application power of the theories they choose to employ in your studies, ensuring that they align with the goals and scope of your projects. (NA?)

  • Apply machine learning techniques to semi-automatically create semantic mappings between ontologies, specifically by calculating the joint distribution of concepts and applying a user-supplied similarity function to generate a similarity matrix between the concepts in the two taxonomies. (NA?)

  • Consider utilizing multiple cross-cutting clusterings in your analysis, as this approach can help identify distinct relationships within structured data and enhance overall understanding of the dataset. (NA?)

  • Pay special attention to the graphical representation of bibliometric maps, utilizing advanced features like zoom functionality, special labeling algorithms, and density metaphors to enhance visualizations, especially when dealing with larger maps. (NA?)

  • Consider leveraging algorithms for identifying similarities between overlapping ontologies from different sources to also identify differences between versions of the same ontology, thereby reducing cognitive load for users and improving overall ontology management efficiency. (NA?)

  • Utilize the FEVER framework to ensure standardized and comparable evaluations of various entity resolution approaches, including both non-learning and learning-based match approaches. (NA?)

  • Carefully consider the granularity of linguistic term sets (LTSs) when conducting qualitative group decision making (QGDM) analyses, as different granularities can impact the accuracy and reliability of results. (NA?)

  • Adopt a triarchic approach to granular computing, incorporating philosophical, methodological, and mechanistic perspectives, while considering both multilevel and multiview understandings of problems. (NA?)

  • Use the “generalized least general generalization” method when attempting to generalize findings across multiple studies or datasets. (NA?)

  • Adopt a broad, inclusive, and judgement-neutral perspective when conducting reviews of cognitive architectures, focusing on the diversity of ideas attempted and your relative success in modeling human cognitive abilities. (NA?)

  • Carefully consider the choice of features, out-of-sample deployment, operating point selection, and potential drastic changes required when evaluating the feasibility and difficulty of recourse in machine learning models. (NA?)

Knowledge Graphs

  • Utilise a debate dynamic framework for fact-checking on knowledge graphs, involving two reinforcement learning agents that extract arguments supporting the thesis or antithesis, and a binary classifier that makes the final judgement based on these arguments. (Hildebrandt et al. 2020)

  • Develop a comprehensive end-to-end solution for product knowledge collection, covering components from ontology construction and enrichment, to data extraction, cleaning, and normalization, while utilizing advanced techniques such as Graph Neural Networks (GNN), transformer, and multi-task learning to efficiently handle the complexity and sparsity of structured data in the retail domain. (X. L. Dong et al. 2020)

  • Consider combining both lexical and statistical information from category instances to derive high-quality axioms from categories, instead of relying solely on either lexical or statistical information. (Heist and Paulheim 2019)

  • Combine symbolic-based methods with walk-based reinforcement learning models to enhance the performance of knowledge graph reasoning tasks while maintaining interpretability. (R. Das et al. 2017)

  • Utilize probabilistic soft logic (PSL) for knowledge graph identification, which enables them to effectively manage large-scale data sets, reason jointly about candidate facts and your associated extraction confidences, identify coherent entities, and incorporate ontological constraints. (NA?)

  • Utilize statistical relational learning (SRL) techniques when working with large-scale knowledge graphs, as these methods allow for efficient handling of sparse relationships within the graph structure. (NA?)

Semantic Web Technologies

  • Utilise a semantic web technology approach to create a unified language for representing heterogenous knowledge about TinyML components, which can then be centralised in a Knowledge Graph (KG) for efficient discovery, interoperation, and management of TinyML systems. (Corneliou et al. 2021)

  • Utilize a combination of distributional semantics and ConceptNet, specifically through the creation of a hybrid semantic space called “ConceptNet Numberbatch,” to achieve superior performance in natural language processing tasks such as word relatedness and solving SAT-style analogies. (Speer et al. 2016)

Commonsense Reasoning

  • Consider incorporating large-scale commonsense knowledge bases, such as ConceptNet, into your studies to enhance the depth and accuracy of your textual analysis and improve overall understanding. (NA?)

Time Series Analysis And Forecasting

  • Incorporate causal inference into your time series predictions to enhance the interpretability and robustness of your models, particularly in the context of medium- and long-term load forecasting for power plants. (K. Yang and Shi 2023)

  • Use the warpDLM framework for analyzing time series of counts, as it combines the advantages of traditional DLMs with the ability to handle discrete data features like zero-inflation, over/under-dispersion, boundedness, and censoring, while providing exact, coherent, and recursive updates for filtering, smoothing, and forecasting distributions. (B. King and Kowal 2023)

  • Utilize a variety of statistical, machine learning, and neural network methods along with scale-free and percentage error-based accuracy metrics to effectively analyze and make predictions about multiple time series data in Internet of Things (IoT) applications. (Tzagkarakis et al. 2022)

  • Utilize the ensemble Kalman filter (EnKF) to accelerate pseudo-marginal MCMC for state space models, thereby reducing computational costs while maintaining reasonable accuracy. (Drovandi et al. 2022)

  • Focus on developing a comprehensive anomaly detection system that incorporates timely, scalable, and robust approaches to accurately identify and respond to anomalies within time series data. (Yue Lu et al. 2022)

  • Employ a Multi-granularity Residual Learning Framework (MRLF) for time-series prediction tasks, which effectively explores multi-granularity patterns by proposing a cross-granularity residual learning net and a Multi-Granularity Confidence Estimator to determine the relevance of specific granularity data for final predictions. (M. Hou et al. 2022)

  • Prioritize developing online algorithms for time series decomposition that can efficiently process high volumes of data with long seasonalities, thereby reducing computational costs and improving overall performance. (Abhinav Mishra, Sriharsha, and Zhong 2022)

  • Employ Distinct Filter Generation Network (DFGN) and Dynamic Adjacency Matrix Generation Network (DAMGN) plugins to improve the accuracy of time series forecasting models by effectively capturing distinct temporal dynamics among entities and dynamic entity correlations, thereby reducing the total number of parameters. (Cirstea et al. 2021)

  • Utilise the Relational Events Model with Spurious Events (REMSE) to control for potential biases arising from spurious events in relational event data, ensuring more accurate and reliable inferences. (Fritz et al. 2021)

  • Utilize a Bayesian approach when dealing with time-varying conditional heteroscedasticity models, specifically through the implementation of a computationally efficient MCMC algorithm based on Hamiltonian Monte Carlo (HMC) sampling. (Karmakar and Roy 2021)

  • Utilise the Case-crossover APriori (CAP) algorithm to provide association and causal rules explaining the occurrences of flooding events, and the Case-crossover APriori Predictive algorithms (CAPP1 and CAPP2) to predict them. (Dhaou et al. 2021)

  • Consider using DeepMVI, a deep learning method specifically tailored for missing value imputation in multidimensional time-series datasets, which outperforms existing methods in terms of accuracy while providing significant improvements in downstream analytics. (P. Bansal, Deshpande, and Sarawagi 2021)

  • Carefully consider the potential effects of sampling and approximate aggregations on model fitting and subsequent forecasts when working with high-dimensional time-series data. (Shuyuan Yan et al. 2021)

  • Consider using the ASMODEE algorithm for detecting ongoing changes in COVID-19 incidence patterns, as it employs a flexible time series framework using a variety of models including linear regression, generalized linear models (GLMs), or Bayesian regression, and utilizes outlier detection inspired by classical Shewhart control charts to signal recent anomalous data points. (Jombart et al. 2020)

  • Utilize the sktimes new forecasting framework, which offers a comprehensive and user-friendly approach to building, tuning, and evaluating composite machine learning models for time series forecasting.’ (Löning and Király 2020)

  • Use meta-learning techniques to improve the generalizability of your models, particularly in the field of time series forecasting. (Oreshkin et al. 2020)

  • Consider using autoregressive deep learning models combined with conditioned normalizing flows to effectively model multivariate temporal dynamics in time series forecasting, enabling accurate predictions and analysis of interaction effects. (Rasul et al. 2020)

  • Use leave-future-out cross-validation (LFO-CV) instead of leave-one-out cross-validation (LOO-CV) for time series analysis, as LFO-CV accounts for the temporal ordering of data and avoids overly optimistic estimates provided by LOO-CV. (Bürkner, Gabry, and Vehtari 2020)

  • Carefully choose the appropriate imputation algorithm for handling missing values in time series data, considering factors like accuracy, efficiency, and parameterization, as well as the specific characteristics of the data being analyzed. (Khayati, Lerner, et al. 2020)

  • Consider utilizing a semi-automatic labelling tool called Label-Less’, which employs unsupervised anomaly detection and accelerated DTW for robust anomaly similarity search, to minimize labeling overhead and enable the creation of large-scale, high-quality KPI anomaly datasets.’ (N. Zhao et al. 2019)

  • Consider using the Empirical Risk Minimization (ERM) method for hierarchical forecasting instead of the Minimum Trace (MinT) method because it relaxes the unbiasedness assumption, directly minimizes the mean squared forecast errors, and can handle high-dimensional hierarchies effectively. (Taieb and Koo 2019)

  • Use the UEA multivariate time series classification archive to ensure a more rigorous evaluation of newly proposed time series classification algorithms, as it provides a diverse range of datasets and addresses the limitations of previous archives. (Bagnall et al. 2018)

  • Employ a fast search of motifs across all lengths to effectively capture all useful activity information within the data. (Linardi et al. 2018)

  • Carefully handle local variations when developing anomaly detection algorithms for seasonal Key Performance Indicators (KPIs) to ensure your effectiveness. (Haowen Xu et al. 2018)

  • Adopt an analyst-in-the-loop’ approach to forecasting at scale, combining configurable models with analyst-driven performance analysis, to ensure accurate and scalable forecasting.’ (Taylor and Letham 2017)

  • Carefully choose the study area and ensure that occurrence data represents a random sample of suitable conditions in the domain, taking into account potential sampling biases and spatial dependencies. (S. J. Phillips et al. 2017)

  • Utilize the soft-Dynamic Time Warping (soft-DTW) technique for comparing time series data because it provides a differentiable loss function that is suitable for various tasks like averaging, clustering, and prediction while maintaining the benefits of traditional DTW. (Cuturi and Blondel 2017)

  • Use the instance profile (IP) to generate abundant shapelet candidates, followed by efficiently pruning candidates that dont align with the definition of shapelets using a novel distribution-aware Bloom filter (DAHF). (Gamboa 2017)

  • Utilize individual survival distribution (isd) models to generate accurate and comprehensive survival probability distributions for individual patients, allowing for improved decision making and treatment planning. (“Analysis-Ready Standardized TCGA Data from Broad GDAC Firehose 2016_01_28 Run” 2016)

  • Conduct multiple resampling experiments on various datasets to ensure robust and reliable comparison of time series classification algorithms. (Bagnall et al. 2016)

  • Utilize Bayesian conditioning when attempting to numerically homogenize partial differential equations (PDEs) like those presented in the study. By doing so, they can effectively identify accurate basis elements for your models, leading to improved predictions and outcomes. (Owhadi 2015)

  • Consider using a flexible nonlinear model that optimizes quantile regression loss coupled with suitable regularization terms to maintain the consistency of forecasts across hierarchies, rather than relying solely on traditional linear autoregressive models. (Blundell et al. 2015)

  • Consider using a two-sample problem formulation combined with a nearest neighbors method to effectively evaluate the correlation between time series data and event data in the context of incident diagnosis. (Chen Luo et al. 2014)

  • Utilise a functional factor model (FFM) when modelling and forecasting electricity spot prices, as it allows for a separate consideration of the dynamics induced by the variations of the merit order curve and those induced by electricity demand. (Liebl 2013)

  • Adopt a functional dynamic factor model (FDFM) for analyzing yield curve data, which combines elements of dynamic factor analysis and functional data analysis to effectively capture the cross-sectional, dynamic, and cross-correlated aspects of the data. (Hays, Shen, and Huang 2012)

  • Employ graphical time series models based on the block-recursive Granger-causal Markov property to effectively capture and analyze dynamic relationships among variables in multivariate time series. (Eichler 2011)

  • Consider using a non-dominated family of mutually singular measures when studying backward stochastic differential equations (BSDEs) and your connections to partial differential equations (PDEs), as this approach allows for greater flexibility and robustness in modeling complex systems. (Soner, Touzi, and Zhang 2011)

  • Consider using a two-step approach to approximate h-step ahead density forecasts, where the first step involves modeling the dynamics of the conditional mean and variance using a Gaussian model, followed by a second step that assumes the h-step ahead density can be approximated by a parametric function characterized by a location parameter and a scale parameter. (Lau and McSharry 2010)

  • Consider using the proposed (_{1}) trend filtering method instead of traditional Hodrick-Prescott (H-P) filtering for estimating underlying trends in time series data, particularly when the underlying trend is expected to be piecewise linear. (S.-J. Kim et al. 2009)

  • Focus on understanding the relationship between the Brownian loop measure and the Brownian bubble measure, as they provide insight into the behavior of Brownian paths within specific domains. (Lawler and Werner 2004)

  • Use the Fourier transform of the solutions to the time-fractional telegraph equation to represent its inverse in terms of stable densities, allowing them to analyze the distribution of a telegraph process with Brownian time. (Orsingher and Beghin 2003)

  • Establish necessary and sufficient conditions for the existence and uniqueness of solutions to linear stochastic evolution equations driven by infinite-dimensional fractional Brownian motion, while considering separately the cases of Hurst parameter above and below 1/2. (Tindel, Tudor, and Viens 2003)

  • Utilise dyadic approximations to construct a canonical geometric rough path associated with a fractional Brownian motion with Hurst parameter greater than 1/4. (Coutin and Qian 2002)

  • Carefully choose appropriate inner product spaces for defining integrals with respect to fractional Brownian motion, considering factors like completeness and density of elementary functions within the chosen space. (Pipiras and Taqqu 2000)

  • Utilize wavelet shrinkage methods for minimax estimation in situations involving spatially variable functions, as these methods offer superior performance compared to traditional linear methods in such scenarios. (Donoho and Johnstone 1998)

  • Utilize wavelet shrinkage methods for minimax estimation in situations involving spatially variable functions, as these methods offer superior performance compared to traditional linear methods in such scenarios. (Donoho and Johnstone 1998)

  • Utilize hierarchical Bayesian time series models to effectively decompose complex joint probability distributions into simpler conditional probabilities, allowing for improved understanding and prediction of temporal processes. (“Maximum Entropy and Bayesian Methods” 1996)

  • Utilize a new method of compactness to establish the existence of martingale solutions and stationary solutions for stochastic Navier-Stokes equations under broad hypotheses about the diffusion term. (Flandoli and Gatarek 1995)

  • Utilize the SureShrink’ methodology, which involves a combination of discrete wavelet transformation, soft thresholding of noisy wavelet coefficients, and Stein’s unbiased estimate of risk for threshold choice, to effectively recover a function of unknown smoothness from noisy, sampled data.’ (Donoho and Johnstone 1995)

  • Use the measure of one-way effect, M_y->x, to quantify the impact of one time series on another, rather than relying solely on traditional methods like Granger causality or correlation analysis. (Hosoya 1991)

  • Consider using self-similar processes with independent increments when studying phenomena where stationarity assumptions may not hold, as these processes provide valuable insights into the underlying structure and behavior of complex systems. (Sato 1991)

  • Consider utilizing the analytic expressions for the infinitesimal generators of the processes related to two-sided Brownian motion with a parabolic drift in terms of Airy functions to develop asymptotics for the global behavior of a wide range of isotonic estimators, including those derived under order restrictions. (Groeneboom 1989)

  • Focus on deriving stochastic partial differential equations for measure-valued branching diffusions and Fleming-Viot diffusion models when the basic space is \(R^1\) and the drift operator is a fractional Laplacian of order \(1 < α <= 2\), as this leads to novel insights into your behavior. (Konno and Shiga 1988)

  • Consider using a weighted occupation time approach for studying measure-valued stochastic processes, as it provides a physically meaningful way to analyze the behavior of these complex systems over time. (Iscoe 1986)

  • Incorporate a multi-scale framework into your time series forecasting models, enabling iterative refinements at different temporal scales, and employing cross-scale normalization to prevent distribution shifts between intermediate forecasts. (Huber 1964)

  • Utilise the Unscented Kalman Filter (UKF) instead of the Extended Kalman Filter (EKF) for nonlinear estimation tasks because it offers better accuracy while maintaining similar computational complexity. (NA?)

  • Utilize wavelet threshold estimators for density estimation due to your ability to achieve nearly optimal performance across various global error measures and function spaces, including the important special cases of invariance and mathematical simplicity. (NA?)

  • Focus on developing specialized algorithms for time series classification tasks, specifically those that can effectively capture and utilize temporal relationships within the data. (NA?)

  • Consider using adaptive parameters in your support vector machines (SVMs) to improve generalization performance and obtain sparser solutions in financial forecasting tasks. (NA?)

  • Utilize numerosity reduction techniques to optimize the efficiency of one-nearest-neighbor Dynamic Time Warping (1NN-DTW) algorithms for time series classification tasks, without compromising accuracy. (NA?)

  • Utilize the SAX (Symbolic Aggregate approXimation) method when working with time series data, as it enables both dimensionality reduction and the definition of distance measures that lower bound corresponding distance measures on the original data, thus improving the efficiency and accuracy of data mining algorithms. (NA?)

  • Critically examine the implications of using big data and smart urbanism in city planning and governance, considering factors like the politics of big urban data, technocratic governance, corporatization of city governance, vulnerabilities of digital infrastructure, and the potential for creating a panoptic city. (NA?)

  • Utilize machine learning techniques to analyze acoustic signals from laboratory earthquake simulations, as this approach can effectively predict the time remaining before a fault fails with high accuracy. (NA?)

  • Conduct repeated resamples of your data to avoid overinterpreting results due to small numerical differences or biases from anomalous data sets. (NA?)

  • Utilise the coherence parameter as a model reduction criterion in kernel-based algorithms for time series prediction, thereby eliminating the need for computationally intensive sparsification procedures. (NA?)

  • Consider using the proposed temporal minimum redundancy - maximum relevance (TMRMR) feature selection approach for handling multivariate temporal gene expression data without losing valuable temporal information during data flattening. (NA?)

  • Utilize diverse machine learning algorithms including Autoregressive Integrated Moving Average (ARIMA), Cubist Regression (CUBIST), Random Forest (RF), Ridge Regression (RIDGE), Support Vector Regression (SVR), and Stacking Ensemble Learning to accurately forecast COVID-19 cumulative confirmed cases in Brazil. (NA?)

  • Consider utilizing deep learning techniques like InceptionTime for time series classification due to its superior accuracy and scalability compared to existing methods. (NA?)

  • Consider employing multiple machine learning techniques such as linear regression, multilayer perceptron, and vector autoregression to improve the accuracy of your COVID-19 forecasts. (NA?)

Granger Causality

  • Utilize the Phase Slope Index (PSI) instead of traditional Granger Causality when attempting to determine the direction of information flow in complex physical systems, as the PSI is less susceptible to errors caused by the presence of multiple independent sources. (NA?)

Natural Language Processing

Word Embeddings

  • Focus on developing new approaches that take into account category features of matched Chinese words, as this will help to effectively capture the relationship between words and improve Chinese Named Entity Recognition (NER) performance. (Qiang He et al. 2023)

  • Consider using distributed representation instead of traditional symbolic representation for natural language processing tasks because it addresses issues like data sparsity, allows for multi-grained semantic representation, and facilitates integration of external knowledge. (N. Ding et al. 2022)

  • Consider utilizing context-dependent embeddings for cross-lingual dependency parsing, as they offer richer semantic and syntactic representations than traditional context-independent word embeddings. (Aldarmaki and Diab 2019)

  • Utilise a combination of hyperbolic embeddings and Hearst patterns to effectively infer concept hierarchies from large text corpora. This approach offers benefits such as improved taxonomic consistency, efficiency, interpretability, and state-of-the-art performance on various benchmarks. (M. Le et al. 2019)

  • Utilize a novel “edge probing” framework when carrying out experiments, allowing them to apply a uniform set of metrics and architectures across multiple tasks. (Tenney et al. 2019)

  • Develop a fully unsupervised framework for learning Multilingual Word Embeddings (MWEs) that directly exploits the relations between all language pairs, rather than relying solely on independently trained Unsupervised Bilingual Word Embeddings (UBWEs). (Xilun Chen and Cardie 2018)

  • Develop multiple models for legal document retrieval, including ones that incorporate deep learning and semantic similarity measures, and test them against a gold standard to identify the best performing model for improved accuracy. (Sugathadasa et al. 2018)

  • Utilize exponential family embeddings (Ef-emb) to create product, trip, and customer embeddings, enabling accurate predictions of consumer behavior and effective personalized marketing strategies. (Behera et al. 2017)

  • Focus on the quality rather than the quantity of the training corpus when developing word embeddings, as evidenced by the superior performance of the smaller Wikipedia dataset (2.1 billion tokens) compared to the larger Google Freebase dataset (100 billion tokens) in producing high-quality vector representations. (Sherkat and Milios 2017)

  • Be aware of potential biases in natural language processing tools, as they can absorb and reproduce human-like semantic biases from the language corpora they are trained on. (Caliskan, Bryson, and Narayanan 2017)

  • Consider using fastText, a linear model with a rank constraint and a fast loss approximation, for efficient and accurate text classification on large datasets. (Joulin, Grave, Bojanowski, and Mikolov 2016)

  • Utilize sparse overcomplete word vector representations because they provide improved interpretability and performance compared to traditional dense word vectors, making them ideal for natural language processing tasks. (Faruqui et al. 2015)

  • Utilise a convolutional neural network to create continuous representations for textual relations, thereby enhancing overall performance on link prediction tasks, especially for entity pairs that have textual mentions. (Gormley, Yu, and Dredze 2015)

  • Consider extending the Skip-gram model to learn multiple embeddings per word type, allowing for improved performance in downstream tasks by accounting for polysemy and homonymy. (Neelakantan et al. 2015)

  • Consider incorporating multi-sense embeddings into your language understanding models, as they have demonstrated improved performance in certain tasks such as part-of-speech tagging, semantic relation identification, and semantic relatedness, despite not showing improvement in others like named entity recognition and sentiment analysis. (Jiwei Li and Jurafsky 2015)

  • Utilize weak supervision techniques, such as automatic generation of questions from knowledge bases and collaborative marking of question paraphrases, to effectively train embedding-based models for open-domain question answering. (Bordes, Weston, and Usunier 2014)

  • Utilise the Gromov-Wasserstein distance to learn correspondences between word embedding spaces in a fully-unsupervised manner, leading to a theoretically-motivated optimization problem that can be solved efficiently, robustly, in a single step, and requires no post-processing or heuristic adjustments. (Dinu, Lazaridou, and Baroni 2014)

  • Utilize the Skip-gram model for learning high-quality distributed vector representations of words and phrases, incorporating techniques like negative sampling and subsampling of frequent words to enhance efficiency and accuracy. (Mikolov, Sutskever, et al. 2013)

  • Utilize the Skip-gram model, which efficiently learns high-quality vector representations of words from large amounts of unstructured text data, and apply extensions like sub-sampling of frequent words and negative sampling to enhance the quality of the vectors and increase training speed. (A. Mnih and Teh 2012)

  • Utilize a novel neural network architecture to effectively embed various symbolic representations found in Knowledge Bases (KBs) into a more flexible continuous vector space, thereby preserving and enhancing the original knowledge and allowing for easy application in modern machine learning techniques for prediction and information retrieval. (Bordes et al. 2011)

  • Combine unsupervised and supervised techniques to effectively learn word vectors that capture both semantic term-document information and rich sentiment content, leveraging both continuous and multi-dimensional sentiment information as well as non-sentiment annotations. (Alm, Roth, and Sproat 2005)

  • Utilise an unsupervised algorithm, DIRT, for automatic discovery of inference rules from text, based on an extended version of Harris Distributional Hypothesis applied to paths in dependency trees of a parsed corpus. (NA?)

  • Utilise an unsupervised algorithm, DIRT, for automatic discovery of inference rules from text, based on an extended version of Harris Distributional Hypothesis applied to paths in dependency trees of a parsed corpus. (NA?)

  • Consider using the Word Movers Distance (WMD) metric for measuring the distance between text documents, as it effectively incorporates semantic similarity between individual word pairs and demonstrates superior performance compared to traditional bag-of-words and term frequency-inverse document frequency methods.’ (NA?)

  • Focus on developing global log-bilinear regression models for word vector representation, as they combine the strengths of global matrix factorization and local context window methods, leading to improved performance in word analogy, similarity, and named entity recognition tasks. (NA?)

  • Utilise RDF2Vec, an innovative technique that employs language modelling approaches to generate latent numerical representations of entities in RDF graphs, thereby facilitating effective data mining tasks. (NA?)

  • Consider combining weak supervision and deep representation techniques to enhance clinical text classification, reducing human effort required for labeled data creation and feature engineering. (NA?)

Sentiment Analysis

  • Use a Composition-based Heterogeneous Graph Multi-channel Attention Network (CHGMAN) to effectively encode a constructed heterogeneous graph, thereby enabling accurate predictions of sentiment polarity in multi-aspect multi-sentiment situations. (X. Song et al. 2024)

  • Consider utilizing a combination of location-based, vocabulary-based, and language detection techniques to ensure accurate and comprehensive data collection for sentiment analysis tasks in African languages. (Muhammad et al. 2023)

  • Aim to create large-scale, diverse, longitudinal, multilingual, and unbiased datasets like Mega-COV to enable comprehensive studies of complex phenomena such as the COVID-19 pandemic. (Abdul-Mageed et al. 2020)

  • Utilize Graph Convolutional Networks (GCNs) when dealing with aspect-based sentiment classification tasks, as they are effective in handling syntactic dependencies and long-range multi-word relations, thereby improving overall performance. (Chen Zhang, Li, and Song 2019)

  • Carefully select appropriate sentiment aggregation measures (such as BullR, BI, VA, and AG) depending on the specific stock market variable being predicted (e.g., returns, volatility, trading volume, or survey sentiment values), and ensure rigorous evaluation through techniques like the Diebold-Mariano test and multiple regression models. (Huina Mao, Counts, and Bollen 2011)

  • Consider incorporating social network information into your sentiment analysis models, as it can lead to statistically significant improvements in user-level sentiment classification accuracy. (Chenhao Tan et al. 2011)

  • Consider utilizing a Bayesian Spatial Following model to analyze Twitter data, assuming homophily in social networks, in order to accurately estimate policy positions for both political actors and ordinary users. (Dodds et al. 2011)

  • Carefully evaluate the performance of various machine learning techniques for sentiment analysis tasks, as they may not perform as well as in traditional topic-based categorization, and consider exploring corpus-based techniques instead of relying solely on prior intuitions. (B. Pang, Lee, and Vaithyanathan 2002)

  • Adopt a novel paradigm for sentiment analysis that integrates linguistics, common-sense computing, and machine learning to improve the accuracy of polarity detection by effectively deconstructing natural language text into concepts and opinion targets. (NA?)

  • Use a combination of manual and automated methods to accurately classify and extract relevant information from social media posts during natural disaster events, ensuring better situational awareness and effective decision-making. (NA?)

Text Classification

  • Consider utilizing ChatGPT for zero-shot text classification tasks, particularly automatic genre identification, due to its superior performance over fine-tuned language models like XLM-RoBERTa in certain situations, potentially reducing the need for extensive manual annotation efforts. (Kuzman, Mozetič, and Ljubešić 2023)

  • Consider using prompt-based learning for the argument-to-keypoint mapping task, as it may lead to improved performance when compared to traditional fine-tuning approaches. (Samin, Nikandish, and Chen 2022)

  • Consider incorporating hierarchical information in text classification tasks via tree-based graph neural networks, as it provides richer structural insights and leads to improved performance compared to existing methods. (Chong Zhang et al. 2021)

  • Distinguish between “intended” and “perceived” sarcasm when developing models for sarcasm detection, as these two forms may require different approaches due to potential socio-cultural differences between authors and readers. (Oprea and Magdy 2019)

  • Carefully choose and report text preprocessing methods when evaluating or comparing different natural language processing models, as even minor changes in preprocessing can significantly impact model performance. (Camacho-Collados and Pilehvar 2017)

  • Utilise machine learning and natural language processing algorithms to automatically retrieve and analyse synthesis parameters from a wide array of materials synthesis journal articles, thereby enabling efficient identification of potential synthesis routes for new materials. (E. Kim et al. 2017)

  • Leverage advanced pre-trained language models like BERT and incorporate an edge-labeling graph neural network within a prototypical network framework to improve the performance of few-shot text classification tasks. (Bruna et al. 2013)

  • Utilise labeled features’, i.e., known relationships between certain input features and classes, to guide the development of discriminative probabilistic models. These ‘labeled features’ can be used to impose soft constraints on the model’s predictions on unlabeled instances, thereby improving the overall performance of the model. (Druck, Mann, and McCallum 2008)

  • Utilize Support Vector Machines (SVMs) for text categorization due to your ability to handle high-dimensional feature spaces, recognize few irrelevant features, and manage sparse instance vectors, leading to improved performance compared to other methods. (NA?)

  • Consider using utility-based evaluation measures instead of traditional ones like recall-precision break-even points, since utility measures provide a more comprehensive and practical approach to evaluating text filtering effectiveness. (NA?)

  • Carefully consider the impact of class skew on feature selection metrics, as the newly proposed “Bi-Normal Separation” (BNS) metric significantly outperforms existing ones in high-skew scenarios, leading to improved accuracy, F-measure, and recall in text classification tasks. (NA?)

  • Consider using Support Vector Machines (SVM) with a tree kernel function to effectively capture the syntactic structures of questions in question classification tasks, leading to improved performance compared to other machine learning methods. (NA?)

  • Consider combining multiple techniques, such as regular expressions and vector space models, to achieve improved accuracy in text categorization tasks. (NA?)

  • Carefully consider the advantages and limitations of local association analysis and global association analysis when developing text-mining strategies for extracting protein-protein interactions from scientific literature. (NA?)

  • Consider various feature selection and transformation methods, including gini index, information gain, mutual information, chi-squared statistics, and supervised latent semantic indexing, to improve the performance of text classification models. (NA?)

  • Create a comprehensive, manually annotated corpus containing both pharmacokinetic (PK) and pharmacodynamic (PD) drug-drug interactions (DDIs) to improve the accuracy and reliability of natural language processing (NLP) techniques in identifying and classifying pharmacological substances and detecting DDIs within biomedical literature. (NA?)

  • Utilize machine learning techniques to accurately calculate the helpfulness of online consumer reviews, thus mitigating the Matthew and ratchet effects that hinder accurate assessment. (NA?)

Named Entity Recognition

  • Consider employing a task-specific prompt framework when working with GPT models for clinical named entity recognition tasks, as it significantly enhances your feasibility for potential clinical applications. (Yan Hu et al. 2023)

  • Consider the impact of multiple factors including language, time period, document type, and annotation tag sets when developing named entity processing systems for historical documents. (“Advances in Information Retrieval” 2022)

  • Consider both the models architecture and the characteristics of the task and dataset when evaluating the generalization behavior of neural network-based models, especially in natural language processing tasks like named entity recognition.’ (Baluja and Fischer 2017)

  • Utilize a combination of internal and external evidences in your named entity recognition (NER) systems, including simple deterministic internal features like capitalization and digitization, internal semantic features of important triggers, internal gazetteer features, and external macro context features. By integrating these different types of evidence, researchers can achieve improved accuracy in recognizing and classifying names, times, and numerical quantities in text. (NA?)

  • Consider integrating multiple evidential features, including word formation pattern, morphological pattern, part-of-speech, head noun trigger, special verb trigger, and name alias feature, through a hidden Markov model (HMM) and a HMM-based named entity recognizer, along with a k-Nearest Neighbor (k-NN) algorithm to address data sparsity, in order to effectively capture local context dependencies and improve the performance of biomedical named (NA?)

  • Utilize Conditional Random Fields (CRFs) for semantic relation extraction (SRE) in biomedical texts, specifically employing either cascaded CRFs or one-step CRFs depending on the availability of prior entity information, leading to improved accuracy and efficiency in identifying and categorizing relationships between entities. (NA?)

  • Utilize the Expectation Maximization (EM) algorithm for inferring ground truth based on team submissions, allowing effective detection of good team performance without reliance on human annotations. (NA?)

  • Consider combining multiple machine learning algorithms and diverse feature sets to achieve optimal performance in named entity recognition tasks within clinical texts. (NA?)

  • Utilize pairwise learning to rank (pLTR) for disease name normalization (DNorm) in biomedical texts, as it enables accurate identification of disease mentions and assignment of unique identifiers, improving overall performance in comparison to traditional lexical normalization and matching techniques. (NA?)

  • Focus on improving the accuracy of chemical entity recognition in text by utilizing advanced techniques such as machine learning algorithms, chemistry and drug lexica, and domain-specific rules, ultimately leading to better identification of chemical compounds and your properties. (NA?)

  • Leverage meta-learning-based continuous cue adjustment methods for few-shot named entity recognition in the electric power domain, using a generative pre-trained language model and a vector of learnable parameters to compensate for the lack of training data. (NA?)

  • Consider using ContrastNER, a prompt-based NER framework that combines discrete and continuous tokens in prompts and employs a contrastive learning approach to improve entity recognition accuracy in low-resource situations without relying on extensive manual prompt engineering and verbalizer design. (NA?)

  • Focus on developing and refining prompt-based strategies to significantly enhance the performance of large language models, such as GPT-3.5 and GPT-4, in processing complex clinical data and extracting meaningful information with minimal training data. (NA?)

Dependency Parsing

  • Utilize bidirectional Long Short Term Memory (BiLSTM) feature representations in combination with traditional parsing techniques to enhance the accuracy of dependency parsing tasks. (Kiperwasser and Goldberg 2016)

Question Answering

  • Utilize specialized language models tailored to the medical domain, even if they are developed in another language, to achieve superior performance in multiple-choice question answering tasks. (Labrak et al. 2023a)

  • Explore innovative directions for machine translation using large language models, such as stylized translation, interactive translation, translation memory-based translation, and a new evaluation paradigm, while addressing privacy concerns and considering future directions like personalized translation and multimodal translation. (C. Lyu, Xu, and Wang 2023)

  • Consider using a combination of shot prompting and context pattern prompting in prompt engineering to improve the performance of large language models in automated medical reporting. (Zandvoort et al. 2023)

  • Utilise a multi-stage fine-tuning approach when working with pretrained transformer models for automatic summarisation of doctor-patient conversations. This involves first summarising segments of the conversation separately before combining and refining these summaries into a complete summary. This methodology allows for the successful handling of long conversations and significantly improves the quality of generated summaries. (T. B. Brown et al. 2020)

  • Consider integrating deep reinforcement learning algorithms within conversational recommendation systems to improve the efficiency and accuracy of the recommendation process. (Yueming Sun and Zhang 2018)

  • Utilize a combination of cross-lingual transfer learning and multilingual training techniques to effectively and rapidly adapt neural machine translation systems to new, low-resourced languages. (Artetxe et al. 2017)

  • Consider implementing Monotonic Chunkwise Attention (MoChA) in your sequence-to-sequence models, as it enables efficient training with standard backpropagation, allows for online and linear-time decoding, and improves performance compared to both soft attention and hard monotonic attention methods. (C.-C. Chiu and Raffel 2017)

  • Utilize end-to-end learning frameworks for task-completion dialogue systems to overcome the limitations of traditional modularized systems, improving overall system performance and robustness to errors. (Xiujun Li et al. 2017)

  • Utilize a semi-automated framework for creating factoid question answering (QA) datasets, which involves generating graph-structured logical forms from a knowledge base and converting them into questions with specific characteristics, allowing for fine-grained analyses of QA systems. (Yu Su et al. 2016)

  • Consider combining state-of-the-art recurrent neural networks with a learning approach for inserting constants into the generated programs to achieve higher accuracy in synthesizing programs from natural language. (Desai et al. 2016)

  • Consider incorporating a knowledge graph and graph neural networks into your fusion-in-decoder framework for open-domain question answering systems, as it improves accuracy and reduces computational costs. (“Proceedings of the 21st ACM SIGPLAN International Conference on Functional Programming” 2016)

  • Explore the feasibility of performing neural machine translation directly on a sequence of characters without any explicit word segmentation, as this could lead to improved translation efficiency and accuracy. (J. Chung, Cho, and Bengio 2016)

  • Utilise a single Neural Machine Translation (NMT) model to translate between multiple languages, requiring no alterations to the model architecture, but rather introducing an artificial token at the start of the input sentence to specify the target language. (M. Johnson et al. 2016)

  • Consider utilizing a combination of hand-crafted templates, web exploration, and automated filtering techniques to efficiently generate high-quality, domain-relevant questions from a knowledge base. (Linfeng Song and Zhao 2016)

  • Use a neural attention-based model to represent questions dynamically according to different answer aspects, rather than converting them into a fixed vector, and leverage global knowledge inside the underlying KB to improve the precision of answer representation and alleviate the out of vocabulary (OOV) problem. (Yuanzhe Zhang et al. 2016)

  • Carefully evaluate your choice of knowledge graph when developing question answering systems, considering factors like maintenance status, ease of access to infrastructure, and compatibility with existing datasets. (Bordes et al. 2015)

  • Utilise a set of proxy tasks to evaluate reading comprehension through question answering, thereby identifying and rectifying the shortcomings of your systems. (Weston et al. 2015)

  • Explore and compare the effectiveness of different attention mechanisms in neural machine translation (NMT) models, specifically considering both global and local attention approaches, to optimize translation accuracy and efficiency. (T. Luong, Pham, and Manning 2015)

Machine Translation

  • Consider utilising the unsupervised detection of translation direction based on the hypothesis that p(translation|original) > p(original|translation), which has demonstrated effectiveness in experiments involving massively multilingual machine translation models across 20 translation directions. (Wastl, Vamvas, and Sennrich 2024)

  • Consider implementing a data augmentation technique called “switch-entity” (SE) to address potential biases in neural machine translation (NMT) systems related to gender and sentiment in translations involving person names. (Jun Wang, Rubinstein, and Cohn 2022)

  • Consider utilising monolingually derived paraphrases to improve the efficiency and effectiveness of statistical machine translation (SMT) systems, particularly in low-density language settings. (Nakov 2021)

  • Focus on developing a robust and efficient deep learning approach for email subject line generation, utilizing a multi-stage training strategy that combines supervised cross-entropy training and reinforcement learning, and incorporating a customized Email Subject Quality Estimator (ESQE) to optimize performance. (Rui Zhang and Tetreault 2019)

  • Focus on developing more efficient and effective neural networks for natural language processing tasks, specifically by exploring alternative architectures like the proposed Multi-level Community-aware Graph Neural Network (MC-GNN) layer, which can potentially address issues related to over-smoothing and improve overall performance. (Daoyuan Chen et al. 2019)

  • Utilize pretrained massively multilingual “seed models” and continue training on data related to the low-resourced language (LRL) of interest, employing a technique called “similar-language regularization”. This involves jointly training on both the LRL and a similar high-resourced language to avoid overfitting to small LRL data. (Neubig and Hu 2018)

  • Carefully consider the limitations of traditional evaluation metrics for text generation tasks, and explore alternative methods such as extractive evaluation based on information extraction systems to better understand the quality of automated generations. (Wiseman, Shieber, and Rush 2017)

  • Consider increasing the attention window and pre-training your Neural Transducer (NT) model with Listen, Attend and Spell (LAS) to significantly enhance the performance of your NT model. (Sainath et al. 2017)

  • Utilise graph-convolutional networks (GCNs) to effectively incorporate syntactic structure into neural attention-based encoder-decoder models for machine translation, leading to significant improvements in translation accuracy. (Bastings et al. 2017)

  • Consider implementing a coverage mechanism in your Neural Machine Translation (NMT) models to mitigate issues of over-translation and under-translation, thereby improving the overall alignment between source and target sentences. (Z. Tu et al. 2016)

  • Consider applying transfer learning to improve the performance of neural machine translation (NMT) models in low-resource language scenarios, by leveraging pre-trained high-resource language models to initialize and constrain training for low-resource language pairs. (Zoph et al. 2016)

  • Consider implementing a hybrid neural machine translation (NMT) model that combines word-level and character-level processing to improve translation accuracy and reduce the occurrence of unknown words in translated texts. (M.-T. Luong and Manning 2016)

  • Utilise the proposed fine-tuning algorithm alongside novel many-to-one translation strategies within the multi-lingual neural machine translation framework. This combination allows for effective zero-resource machine translation, performing as well as a single-pair neural translation model trained with up to 1 million direct parallel sentences of the same language pair, and surpassing pivot-based translation strategies. (Firat et al. 2016)

  • Consider applying knowledge distillation techniques to reduce the size of neural machine translation models, specifically through the use of sequence-level knowledge distillation, which improves performance and eliminates the need for beam search. (Yoon Kim and Rush 2016)

  • Utilise a deep LSTM network with 8 encoder and 8 decoder layers using residual connections as well as attention connections from the decoder network to the encoder to improve the performance of Neural Machine Translation (NMT) systems. (Yonghui Wu et al. 2016)

  • Carefully consider the impact of domain mismatch, amount of training data, rare words, long sentences, word alignment, and beam search on the performance of neural machine translation systems, as these factors can significantly affect the quality of translations produced. (Bahdanau, Cho, and Bengio 2014)

  • Utilise a meta-learning approach called Meta-MT to efficiently adapt Neural Machine Translation (NMT) systems to various target domains with minimal in-domain data. (Bahdanau, Cho, and Bengio 2014)

  • Carefully examine the limitations of neural machine translation models, particularly your ability to handle long sentences and unknown words, and explore potential improvements through techniques like gated recursive convolutional neural networks. (K. Cho, Merrienboer, Bahdanau, et al. 2014)

  • Consider implementing a technique to handle rare words in Neural Machine Translation (NMT) systems, specifically by training an NMT system on data augmented by the output of a word alignment algorithm, enabling the system to identify the position of corresponding words in the source sentence for each out-of-vocabulary (OOV) word in the target sentence. This information can then be utilised in a post-processing step that translates every OOV word using a (M.-T. Luong et al. 2014)

  • Carefully consider the potential benefits of including linguistic features in neural machine translation models, as they can significantly enhance model performance across multiple evaluation metrics. (NA?)

  • Carefully consider and optimize the choice of batch size, learning rate, warmup steps, maximum sentence length, and checkpoint averaging when training the Transformer sequence-to-sequence model for neural machine translation tasks. (NA?)

  • Adopt an interdisciplinary approach to studying automatic translation from signed to spoken languages, incorporating insights from computer vision, machine translation, and linguistics, and involving deaf and hearing end users in use case identification, data collection, and evaluation processes. (NA?)

Dialogue Systems

  • Adopt a prompt pool method for class-incremental continual learning in dialog state tracking, allowing automatic identification of tasks and selection of appropriate prompts during testing. (Hong Liu et al. 2023)

  • Carefully balance the choice of decoding algorithm and the length of the bots utterances to achieve optimal human judgements of quality in open-domain chatbot development.’ (Roller et al. 2021)

  • Carefully balance the choice of decoding algorithm and the length of the bots utterances to achieve optimal human judgements of quality in open-domain chatbot development.’ (Roller et al. 2020)

  • Consider building and deploying a role-playing game to facilitate lifelong open-domain dialogue learning, allowing models to progressively improve through interactions with human players, leading to more efficient and cost-effective data collection compared to traditional crowdsourced methods. (Shuster et al. 2020)

  • Consider utilizing pretrained language models like GPT-2 for task-oriented dialogue systems, as they can help overcome data scarcity challenges and potentially lead to more engaging and eloquent conversational agents. (Budzianowski and Vulić 2019)

  • Consider utilizing pretrained language models like GPT-2 for task-oriented dialogue systems, as they can help overcome data scarcity challenges and lead to more engaging and eloquent conversational agents. (Wenhu Chen, Chen, et al. 2019)

  • Develop a proactive dialogue system by planning dialogue strategy over a knowledge graph, allowing the model to effectively utilise related knowledge to generate more diverse multi-turn conversations. (Wenquan Wu et al. 2019)

  • Move away from current evaluation metrics for dialogue response generation systems, as they demonstrate weak or no correlation with human judgements, and instead develop new metrics that correlate more strongly with human assessment. (C.-W. Liu et al. 2016)

  • Develop a dialog-conditioned path traversal model called AttnIO’, which uses two directions of attention flows to fully exploit the rich structural information in a knowledge graph (KG) and improve the performance of knowledge selection problems in dialogue systems.’ (K. Cho, Merrienboer, Gulcehre, et al. 2014)

  • Consider using a probabilistic framework for dialog simulation, incorporating separate models for automatic speech recognition, user behavior, and natural language understanding, to optimize strategy learning and improve the efficiency and effectiveness of spoken dialogue systems. (Pietquin and Dutoit 2006)

  • Utilise a combined approach of supervised and reinforcement learning to effectively train dialog systems. Supervised learning is employed to estimate a model of the user, specifically the MDP parameters that quantify the users behaviour. Following this, reinforcement learning is applied to estimate the optimal strategy while the system interacts with the simulated user. (NA?)

  • Consider the multi-dimensional impact of digital resurrection, including rhetoric, everyday experiences, and emotions, when studying the effects of chatbots designed to simulate deceased individuals. (NA?)

Vision And Audio Processing

  • Consider using multiple codebooks and designing specific architectures for multi-code sampling and monotonic alignment when working with real-world speech data for text-to-speech synthesis. (L.-W. Chen, Watanabe, and Rudnicky 2023)

  • Consider utilising the alignment mechanism proposed in RAD-TTS as a generic alignment learning framework for a wide range of neural TTS models, as it enhances alignment convergence speed, simplifies the training pipeline, and improves the perceived speech synthesis quality. (Badlani et al. 2021)

  • Carefully balance the trade-off between fine-tuning parameters and voice quality in custom voice applications, taking into account factors such as memory storage and serving costs. (Mingjian Chen et al. 2021)

  • Develop a more comprehensive degradation model for single image super-resolution tasks, incorporating multiple factors like blur, downsampling, and noise, and allowing for your random shuffling to better capture the diversity of real-life image degradations. (Kai Zhang et al. 2021)

  • Consider using an attentive graph neural network (AGNN) for zero-shot video object segmentation (ZVOS), as it enables efficient information fusion over video graphs, capturing richer and higher-order relations between video frames, resulting in improved accuracy for foreground estimation. (Wenguan Wang et al. 2020)

  • Consider adopting a multi-modal approach to music emotion recognition (MER) by incorporating diverse sources of information such as audio, MIDI, and lyrics, which can potentially lead to significant improvements in the overall performance of MER systems. (Cambria et al. 2020)

  • Consider using 4D radar technology for object detection in autonomous vehicles because it offers superior accuracy and robustness in adverse weather conditions compared to traditional 3D radar systems. (Major et al. 2019)

  • Consider using a proposal-based object detector that allows each proposal to predict a set of correlated instances rather than a single one, especially in crowded scenes. (Z. Cai and Vasconcelos 2019)

  • Focus on developing pretext-invariant representation learning (PIRL) techniques for self-supervised learning from images, as opposed to traditional covariant methods, in order to create more accurate and robust image representations. (Misra and Maaten 2019)

  • Consider developing a novel temporal loss function that accounts for higher time derivatives of point positions and encourages mingling to prevent halos when dealing with temporally coherent feature spaces in point clouds. (Prantl et al. 2019)

  • Focus on developing adaptive masked proxies for few-shot segmentation, which involves creating a normalized masked average pooling layer to generate class signatures from base embeddings, fusing them with previously learned class signatures through multi-resolution imprinting, and updating the weights of previously learned classes without back-propagation to enable sample-efficient learning. (Siam, Oreshkin, and Jagersand 2019)

  • Use a standardized dataset to evaluate and compare the performance of various nonlinear modeling techniques for emulating guitar amplifiers, allowing for more accurate assessments and facilitating advancements in the field. (“145th Audio Engineering Society Convention” 2018)

  • Utilise a strong conditional generative model to sample counterfactual inputs that either change or preserve classifier behaviour, rather than using ad hoc in-filling approaches such as blurring or injecting noise, which generate inputs far from the data distribution and ignore informative relationships between different parts of the image. (C.-H. Chang et al. 2018)

  • Consider incorporating “global style tokens” (GSTs) into your end-to-end speech synthesis systems like Tacotron, as these tokens can effectively model a wide range of acoustic expressiveness without explicit labels, allowing for better control and transfer of speaking style in synthetic speech. (Yuxuan Wang et al. 2018)

  • Consider utilizing a two-branch dense comparison module combined with an iterative optimization module for effective few-shot semantic segmentation, allowing for accurate segmentation of new classes with limited annotated images. (L.-C. Chen et al. 2017)

  • Utilize a modified version of the WaveNet architecture for singing synthesis, focusing on modeling features produced by a parametric vocoder that separates the influence of pitch and timbre, thereby improving efficiency, reducing training and generation times, and providing greater flexibility in matching target melodies. (Blaauw and Bonada 2017)

  • Consider incorporating a light-weight featurized image pyramid network (LFIP) within your single-shot detection framework to effectively address the challenge of detecting very small or large objects, while maintaining high detection accuracy and real-time speed. (Yunpeng Chen et al. 2017)

  • Carefully manipulate the content and style information of input image pairs to create more abundant and robust samples, enhancing the generalization of model training. (DeVries and Taylor 2017)

  • Develop and utilise large-scale, multi-modal video similarity evaluation datasets like Tencent-MVSE to improve the accuracy and efficiency of video recommendation systems. (Abu-El-Haija et al. 2016)

  • Focus on developing fully convolutional networks that are capable of proposing instance-level segment candidates, rather than solely focusing on semantic segmentation. (J. Dai et al. 2016)

  • Consider combining distinct datasets through a hierarchical view of object classification, allowing them to leverage labeled detection images to learn precise localization of objects while utilizing classification images to enhance vocabulary and robustness. (Redmon and Farhadi 2016)

  • Explore combining neural networks and visual semantic embeddings to predict answers to simple questions about images, while also developing a question generation algorithm to produce larger, more balanced datasets for improved model performance. (Foti et al. 2014)

  • Develop a coarse-to-fine compositional representation for temporal grounding tasks, allowing for improved sensitivity to different video and query granularity. (Diederik P. Kingma and Ba 2014)

  • Consider using trainable speaker embeddings for multi-speaker text-to-speech systems, as it enables a single neural TTS system to learn hundreds of unique voices from limited data per speaker, while preserving speaker identities nearly perfectly. (K. Cho, Merrienboer, Gulcehre, et al. 2014)

  • Utilize a Bayesian estimation framework to effectively fuse multi-band images, taking into consideration the unique characteristics of each type of image (such as spectral or spatial response) and employing advanced techniques like Markov Chain Monte Carlo (MCMC) algorithms to accurately estimate the unknown scene. (Q. Wei, Dobigeon, and Tourneret 2013)

  • Develop innovative methods for accurately estimating the shear applied to galaxy images in the presence of various sources of noise and uncertainty, while minimizing bias and maximizing signal-to-noise ratio, in order to improve the understanding of dark matter and dark energy distributions in the universe. (Bridle et al. 2009)

  • Consider exploring alternative sparsity techniques beyond just pruning pre-trained models, as there are potential benefits in terms of improved performance, reduced overfitting, increased robustness, and decreased model complexity. (Janowsky 1989)

  • Focus on developing advanced statistical models and machine learning techniques to improve the accuracy and efficiency of automatic music transcription systems. (NA?)

  • Consider using Generative Adversarial Networks (GANs) for waveform synthesis in speech technology because it enables unrestricted use of feedforward architectures capable of parallel inference, leading to improved efficiency and performance. (NA?)

  • Incorporate an estimated network (Es-Network) within your Tacotron 2 framework to address the over-smoothness issue in speech synthesis, thereby enabling the model to better focus on predicting individual features of mel spectrograms and ultimately resulting in more natural and expressive synthesized speech. (NA?)

  • Consider employing a Tacotron-based multispeaker acoustic model trained on read-only speech data for achieving prosody control at the phoneme level when developing a text-to-rapping/singing system. (NA?)

  • Focus on developing machine learning models that generalize well across diverse image types and experimental variations, even in the presence of dataset biases, to improve the accuracy of nucleus segmentation. (NA?)

  • Consider adopting Latent Diffusion Models (LDMs) for local text-guided image editing tasks, as they provide faster inference times and improved precision compared to traditional diffusion models operating directly in the pixel space. (NA?)

  • Utilise a combination of Voronoi vertices and Delaunay triangulations to accurately approximate a smooth surface from a finite set of sample points. (NA?)

Image Segmentation

  • Consider implementing a Cycle-Resemblance Attention Prototype Network (CRAPNet) for few-shot medical image segmentation tasks, as it effectively preserves spatial correlations between image features and seamlessly integrates them into traditional prototype networks. (H. Ding et al. 2022)

  • Consider using normalizing flows (NFs) in an invertible generative framework to model the distribution mapping of point clouds, as NFs offer several advantages such as transforming complex distributions into disentangled code space, being invertible and lossless, and realizing the encoding and decoding process in a unified framework. (Guerrero et al. 2018)

  • Utilise the Tversky loss function when dealing with highly imbalanced data sets in medical image segmentation, as it provides a better trade-off between precision and recall, leading to improved overall performance. (Salehi, Erdogmus, and Gholipour 2017)

  • Consider using the SegNet architecture for image segmentation tasks because it effectively balances performance and efficiency through its unique decoding technique that utilizes max-pooling indices for non-linear upsampling, thereby improving boundary delineation, reducing the number of parameters, and allowing easy integration into various encoder-decoder architectures. (NA?)

Object Detection

  • Consider implementing a hardware-friendly quantization scheme that avoids the need for floating point arithmetic operations during inference, while identifying and addressing issues related to instability during the fine-tuning stage of the quantization process. (Z. Cai et al. 2017)

  • Carefully consider the speed/accuracy trade-off when choosing a detection architecture for your specific application and platform, taking into account factors such as feature extractors, image resolution, and hardware limitations. (Jonathan Huang et al. 2016)

  • Replace traditional surrogate regression losses like l1 and l2-norms with a metric loss calculated based on Intersection over Union (IoU) for improved performance in 2D object detection tasks. (Kosub 2016)

Scene Understanding

  • Adopt a universal self-supervised 3D scene pre-training framework, named Point-GCC, which effectively leverages geometry and color information via a Siamese network with hierarchical supervision. This approach bridges the gap between pre-training and downstream tasks, leading to significant improvements in performance across a variety of tasks and datasets. (G. Fan et al. 2023)

Speaker Identification

  • Utilize Deep Feature Loss (DFL) for feature-domain supervised denoising in order to optimize the enhancement network in the hidden activation space of a pre-trained auxiliary speaker embedding network, leading to improved speaker verification accuracy in adverse environments. (Kataria et al. 2019)

  • Consider employing a convolutional time-delay deep neural network structure (CT-DNN) for speaker feature learning, as it can produce high-quality speaker features even with a single feature (0.3 seconds including the context), leading to a lower equal error rate (EER) in speaker verification tasks. (Chao Li et al. 2017)

Speech Synthesis

  • Leverage a hierarchical sequence-to-sequence modeling approach for generating high-fidelity music from text descriptions, utilizing a combination of semantic and acoustic tokens derived from pretrained models. (A. Agostinelli et al. 2023)

  • Consider using the M4Singer dataset, a large, high-quality, multi-style, multi-singer Mandarin singing corpus with elaborate musical score annotations, to improve the performance of various singing voice synthesis tasks such as score-based SVS, controllable singing voice, singing voice conversion, and automatic music transcription. (R. Huang et al. 2021)

  • Consider using a modified version of the StyleMelGAN vocoder called Streamwise StyleMelGAN (SSMGAN) for frame-by-frame generation of wideband speech at low delay, with reasonable computational complexity, especially when dealing with streaming applications. (A. Mustafa et al. 2021)

  • Consider using HiFi-GAN, a novel approach combining multi-scale and multi-period discriminators, to achieve both efficient and high-fidelity speech synthesis, as demonstrated by its superior performance in generating 22.05 kHz high-fidelity audio 167.9 times faster than real-time on a single V100 GPU. (J. Kong, Kim, and Bae 2020)

  • Consider integrating the training and conversion processes of speech and singing into one framework, allowing them to leverage normal speech data for singing voice conversion training, thereby improving the robustness of the system, particularly when the singing database is small. (Liqiang Zhang et al. 2020)

  • Consider using a phonemic input representation to encourage sharing of model capacity across languages, and incorporating an adversarial loss term to encourage the model to disentangle its representation of speaker identity from the speech content, in order to achieve successful cross-language voice cloning. (Liqiang Zhang et al. 2019)

  • Carefully evaluate the trade-off between naturalness and similarity of the cloned voices versus the computational resources needed for speaker encoding and adaptation, particularly in low-resource deployments. (S. O. Arik et al. 2018)

  • Consider implementing style tokens into your Tacotron model to enable better prosody control in speech synthesis, allowing for more accurate and expressive speech generation. (Yuxuan Wang et al. 2017)

Quantum Machine Learning

  • Employ an active learning approach to Hamiltonian learning, specifically the Hamiltonian active learning (HAL) algorithm, which enables efficient and accurate estimation of Hamiltonian parameters within a specified learning error through minimal queries. (Dutt et al. 2023)

  • Consider using transformer quantum states (TQS) as a multi-purpose model for quantum many-body problems, as it can generate the entire phase diagram, predict field strengths with experimental measurements, and transfer knowledge to new systems it has never been trained on before, all within a single model. (Y.-H. Zhang and Ventra 2023)

  • Consider utilizing quantum optical neural networks (QONNs) as they can effectively combine the versatility of neural networks with the complexity of quantum optical systems, allowing them to perform a wide range of quantum information processing tasks, including novel protocols such as quantum optical state compression for quantum networking and black-box quantum simulation. (Steinbrecher et al. 2018)

  • Utilize the method of differential equations to derive the master integrals for two-loop Feynman integrals, employing a canonical basis for the integrals, and expressing the results in terms of multiple polylogarithms for optimal numerical evaluation. (Baglio, Ninh, and Weber 2016)

  • Consider utilizing quantum computing techniques for machine learning tasks, particularly those involving large datasets and high-dimensional vectors, due to its potential for exponential speedups over classical algorithms. (S. Lloyd, Mohseni, and Rebentrost 2013)

  • Utilize the Berenstein-Sjamaar theorem, rephrased into a usable form (Theorem 1), as a theoretical foundation for studying the (N)-representability problem, which involves understanding the constraints on the occupation numbers of quantum mechanical systems. (NA?)

  • Utilise the smooth entropy formalism within the field of one-shot information theory to effectively analyse and understand complex systems involving quantum channels. (NA?)

  • Consider employing a learning-based approach using a simple photonic architecture to process information at unprecedented data rates, as demonstrated through the implementation of a semiconductor laser subject to delayed self-feedback and optical data injection to solve computationally hard tasks. (NA?)

  • Utilise random tensor networks as a powerful tool for exploring holographic duality, as they naturally adhere to entanglement area laws and can be interpreted as the partition function of a classical ferromagnetic Ising model, allowing for the calculation of various entropies and the analysis of bulk-boundary correspondences. (NA?)

  • Investigate the potential of quantum computation to enhance machine learning (ML) algorithms, especially for large datasets and computationally difficult problems, despite existing limitations and uncertainty regarding the feasibility of implementing quantum algorithms in practice. (NA?)

  • Consider using a quantum algorithm for linear regression when dealing with large datasets, as it offers significant improvements in efficiency compared to traditional methods. (NA?)

  • Utilize the Steinmann relations when conducting the heptagon cluster bootstrap process. This significantly reduces the computational complexity involved in calculating seven-point amplitudes in planar N = 4 supersymmetric Yang-Mills theory, thereby making higher-loop contributions more computationally accessible. (NA?)

  • Utilise the CFT Froissart-Gribov formula to understand how the spectrum organises into analytic families and gain control over individual OPE coefficients as opposed to averages. (NA?)

  • Consider utilizing deep neural networks over shallow ones when attempting to efficiently represent quantum many-body states, as deep neural networks can effectively capture the majority of physical states, whereas shallow networks like restricted Boltzmann machines struggle to do so without significant computational complexity. (NA?)

  • Consider extending adversarial training to the quantum domain and utilize quantum circuits to construct generative adversarial networks, enabling them to compute gradients and train quantum generative adversarial networks effectively. (NA?)

  • Investigate the potential of quantum computation to enhance the efficiency of machine learning algorithms, especially for large datasets and computationally intensive tasks, despite the current limitations of quantum technology. (NA?)

  • Utilize a representation of any atom in any chemical environment for the creation of effective quantum machine learning (QML) models of common electronic ground-state properties. This representation is founded on scaled distribution functions that take into consideration both elemental and structural degrees of freedom. (NA?)

  • Utilize the Mathematica package PolyLogTools to efficiently explore and analyze the algebraic structures of multiple polylogarithms (MPLs) in various fields of high-energy physics. (NA?)

  • Consider implementing a quantum convolutional neural network (QCNN) for efficient training and implementation on near-term quantum devices, as it uses O(log(N)) variational parameters for input sizes of N qubits. (NA?)

  • Utilize the inherent similarity between quantum computing and kernel methods in machine learning to develop more effective quantum machine learning algorithms. (NA?)

  • Carefully choose appropriate reference methods for generating datasets of excited-state data, considering factors such as accuracy, computational cost, and suitability for various applications. (NA?)

  • Utilize quantum neurons as a foundation for constructing quantum feed-forward neural networks capable of universal quantum computation, while employing fidelity as a cost function for efficient training. (NA?)

  • Carefully examine the relationship between the kernel function and the observable expectation of quantum states to determine the potential for quantum prediction advantage in machine learning tasks. (NA?)

  • Carefully consider the potential impact of noise-induced barren plateaus (NIBPs) on the scalability of Variational Quantum Algorithms (VQAs), as NIBPs can lead to exponential scaling issues and destroy quantum speedup. (NA?)

  • Focus on developing photonic systems capable of being dynamically programmable, scalable to hundreds of modes and photons, and able to access a class of quantum circuits that could not be efficiently simulated by classical hardware. (NA?)

Quantum Computing Basics

  • Consider utilizing the proposed generalized stabilizer formalism when attempting to simulate arbitrary quantum circuits on a classical computer, as it combines the density matrix and stabilizer representations, enabling efficient simulations for special cases where the input state and number of non-Clifford gates are restricted. (Bermejo-Vega and Nest 2012)

Quantum Error Correction

  • Consider multiple errors and encoding of multiple qubits when studying quantum error correction, as some essential properties only emerge in these cases. (Roffe 2019)

Quantum Algorithms

  • Consider extending molecular bootstrap embedding to make it suitable for implementation on a quantum computer, allowing them to solve the electronic structure problem of a large molecule as an optimization problem for a composite Lagrangian governing fragments of the total system, thereby achieving a quadratic speedup over the classical algorithm. (H.-C. Zhou, Long, and Yaghi 2012)

  • Consider pushing standard techniques for speeding up RSA to your extremes, creating a much larger gap between the legitimate users costs and the attacker’s costs, thus providing a reasonable level of concrete security against quantum attacks.’ (Maurer 1995)

  • Carefully consider the dimensionality of your quantum systems, as one-dimensional systems can possess unexpected complexities and limitations compared to higher-dimensional systems. (NA?)

  • Strategically combine classical solvers with quantum processors based on problem complexity to optimize computational efficiency. (NA?)

  • Focus on developing error mitigation techniques for noisy quantum processors, as they can significantly improve your computational capabilities without requiring additional hardware modifications. (NA?)

Quantum Machine Learning Algorithms

  • Focus on developing quantum algorithms for solving linear systems of equations, as they offer significant advantages over classical algorithms, particularly when dealing with large datasets and low condition numbers. (Harrow, Hassidim, and Lloyd 2009)

  • Focus on developing sequential protocols for hidden quantum channel discrimination problems, as they offer significant advantages over non-sequential approaches in terms of achieving perfect discrimination and saturating the Heisenberg limit. (Giovannetti, Lloyd, and Maccone 2004)

  • Utilize the single-valued harmonic polylogarithms (SVHPLs) framework when analyzing the multi-Regge limit of the six-point remainder function in the context of maximally supersymmetric N=4 Yang-Mills theory. (NA?)

  • Utilize the principles of quantum theory, particularly state superposition and quantum parallelism, to develop a novel quantum reinforcement learning (QRL) method that can significantly enhance the efficiency and effectiveness of reinforcement learning algorithms. (NA?)

  • Carefully consider the impact of jet algorithms on factorization and power corrections when studying jet shapes and jet algorithms in SCET. (NA?)

  • Carefully consider the appropriate factorization scheme when analyzing complex systems involving multiple scales, ensuring that the chosen scheme accurately reflects the underlying physics and allows for meaningful interpretation of the results. (NA?)

  • Use neural-network quantum states (NQS) to accurately represent and analyze complex many-body quantum systems, leveraging reinforcement learning techniques to optimize the network parameters. (NA?)

  • Avoid using random initialization in parametric circuit approaches for quantum simulations, as it can lead to exponentially small gradients and poor performance. (NA?)

  • Explore the potential of active learning machine systems in creating new quantum experiments, as demonstrated by the studys application of projective simulation models to design complex photonic quantum experiments that produce high-dimensional entangled multiphoton states.’ (NA?)

  • Utilize an adaptive variational algorithm for precise molecular simulations on a quantum computer, which grows the ansatz systematically one operator at a time according to the specific molecule being studied, resulting in a small number of parameters and shallow-depth circuits. (NA?)

  • Utilise Quantum Imaginary Time Evolution (QITE) as a powerful tool for studying complex quantum systems. It offers significant advantages over traditional methods, providing exponential reductions in space and time costs, making it particularly suitable for near-term quantum devices. (NA?)

  • Utilise the qBAS score as a metric for evaluating the performance of quantum circuits in generating uniform patterns across various datasets, ensuring a balance between precision and recall. (NA?)

  • Consider using a variational algorithm for simulating imaginary time evolution on a hybrid quantum computer, as it allows for efficient representation of many-body quantum states and can be implemented with current quantum computers. (NA?)

  • Consider using quantum generative adversarial networks (qGANs) for accurate and efficient distribution learning, particularly for multi-modal data, by combining a quantum generator with a classical discriminator to optimize the generators parameters through alternating update steps.’ (NA?)

  • Utilise the parameter shift rule for estimating gradients of expectation values of quantum measurements, which allows them to estimate gradients using the same or nearly the same architecture that executes the original circuit. (NA?)

  • Focus on utilizing singular value transformation as a unifying framework for understanding and optimizing quantum algorithms across various domains, leading to significant performance improvements. (NA?)

  • Consider utilising PauliNet, a deep-learning wave function ansatz, to achieve nearly exact solutions of the electronic Schrodinger equation, as it outperforms comparable state-of-the-art VMC ansatzes for atoms, diatomic molecules and a strongly-correlated hydrogen chain while remaining computationally efficient. (NA?)

  • Adopt a co-design approach when working with quantum algorithms, developing the expression of algorithms alongside the hardware itself to optimize execution. (NA?)

  • Utilize a circuit-centric design approach for developing quantum algorithms for supervised learning tasks, which involves understanding a generic strongly entangling quantum circuit as the core of the machine learning model. (NA?)

  • Optimize the hyperparameters of your chosen representation to achieve a compact yet informative input representation, enabling faster and more accurate machine learning algorithms. (NA?)

  • Carefully choose the cost function in variational quantum algorithms (VQAs) to avoid exponentially vanishing gradients (barren plateaus) and ensure trainability, particularly for global cost functions used in abstract applications. (NA?)

  • Focus on developing scalable architectures for quantum annealing and reducing circuit complexity in digital quantum algorithms for optimization such as QAOA, while taking into account the limitations imposed by the qubit connectivity of most hardware platforms. (NA?)

  • Carefully consider the conditions under which quantum algorithms can offer provably efficient solutions for large-scale machine-learning models, specifically when the models are sufficiently dissipative and sparse, with small learning rates. (NA?)

Quantum Neural Networks

  • Combine quantum computation with classical neural network theory to produce a quantum computational learning algorithm, leading to a quantum associative memory with a capacity exponential in the number of neurons. (D. Ventura and Martinez 1998)

  • Focus on developing a Quantum Neural Network (QNN) model that combines the nonlinear, dissipative dynamics of neural computing with the linear, unitary dynamics of quantum computing, while satisfying the three minimum requirements for a meaningful QNN based on the Hopfield Neural Network model and containing the feature of associative memory. (NA?)

  • Aim to achieve exponential localization of isolated Majorana modes at wire ends and demonstrate non-Abelian braiding behavior in order to establish the viability of topological quantum computing in solid-state systems. (NA?)

  • Utilise hierarchical quantum circuits for performing binary classification of classical data encoded in a quantum state, as these circuits provide increased expressivity and accuracy, particularly when dealing with highly entangled quantum states. (NA?)

  • Utilise the proposed quantum Boltzmann machine (QBM) for machine learning tasks, which involves replacing classical spins or bits with quantum bits (qubits) in the Hamiltonian equation, allowing for more complex computations and potentially better performance compared to traditional methods. (NA?)

  • Utilize deep neural network generative models for efficient and scalable density matrix reconstruction in order to overcome the curse of dimensionality inherent in describing quantum states. (NA?)

Quantum Annealing

  • Consider using quantum annealing (QA) for machine learning problems, particularly when dealing with small datasets, as it offers performance advantages compared to classical computational approaches. (NA?)

Ethics And Fairness In Machine Learning

  • Develop a comprehensive framework for evaluating the trustworthiness of large language models (LLMs) by considering eight key aspects - truthfulness, safety, fairness, robustness, privacy, machine ethics, transparency, and accountability - and conducting rigorous testing using diverse datasets and evaluation metrics. (L. Sun et al. 2024)

  • Incorporate ethical considerations, such as transparency, bias mitigation, privacy protection, risk assessment, accountability, continuous monitoring, ethical decision-making, human oversight, recruiting data science experts, and developing best practices, when integrating ChatGPT into marketing practices. (Rivas and Zhao 2023)

  • Prioritize developing methods to detect and mitigate privacy leaks, biases, toxicity, misinformation, and intellectual property infringements in AI-Generated Content (AIGC) models, ensuring responsible and ethical use of these technologies. (L.-W. Chen, Watanabe, and Rudnicky 2023)

  • Focus on developing a collaborative, multidisciplinary approach to identify, quantify, and mitigate biases in large language models, considering the ethical implications and the need for equity, transparency, and responsibility in AI systems. (Ferrara 2023)

  • Maintain vigilance and implement rigorous fact-checking and verification processes while utilizing large language models like ChatGPT, ensuring transparency and accountability in your work. (Dis et al. 2023)

  • Consider the impact of intelligent machines on all three Darwinian properties of culture - variation, transmission, and selection - as they investigate the emerging phenomenon of machine culture. (Brinkmann et al. 2023)

  • Focus on developing strategies to detect and mitigate the risks associated with the misuse of increasingly powerful AI language models, including identifying plagiarism, stopping responses to maliciously intended queries, assigning responsibility for damages caused, ensuring distinction between facts and fictions, and promoting ethical development of AI technology. (J. Chatterjee and Dethlefs 2023)

  • Ensure that AI systems are designed and deployed in a way that supports citizens basic liberties, promotes fair equality of opportunity, provides the greatest benefit to those who are worst-off, and aligns with the institutions and values required by justice.’ (Gabriel 2022)

  • Prioritize allocating significant portions of your AI R&D budgets towards ensuring safety and ethical use, focusing on solving technical challenges related to oversight, robustness, interpretability, risk evaluations, and addressing emerging challenges, while simultaneously advocating for the establishment of national and international governance frameworks to monitor and regulate AI development. (Jumper et al. 2021)

  • Carefully balance multiple fairness criteria while considering detection performance, as enforcing just one fairness criterion might not guarantee fairness in other aspects. (Burkholder et al. 2021)

  • Aim to develop a principle-based approach to AI alignment that combines various elements such as instructions, intentions, revealed preferences, ideal preferences, interests, and values in a systematic way, taking into account the potential impact of AI on human lives and societies. (Gabriel 2020)

  • Adopt stricter empirical standards, such as thorough hyperparameter tuning, sliced analysis, ablation studies, sanity checks, reporting negative results, sharing experimental records, utilizing alternative paper formats, increasing conference page limits, promoting collaboration, providing author contributions, improving reviewer tools, and expanding venue options, to enhance the overall rigor and credibility of machine learning research. (Real et al. 2018)

  • Adopt an interdisciplinary approach to study machine behavior, combining knowledge from various scientific disciplines like computer science, mathematics, engineering, neuroscience, collective behavior, and social theory, and employing randomized experiments, observational inference, and population-based descriptive statistics to define measures of micro and macro outcomes. (K. Patel 2017)

  • Recognize and analyze the value-ladenness of algorithms, considering your potential to create moral consequences, reinforce or undercut ethical principles, and enable or diminish stakeholder rights and dignity. (NA?)

  • Ensure that your tool designs are informed by an understanding of practitioners actual challenges and needs for support in developing fairer ML systems, rather than being solely driven by the availability of algorithmic methods.’ (NA?)

  • Carefully consider the trade-offs between individual and group notions of fairness, and recognize that different fairness criteria reflect different underlying value systems. (NA?)

  • Focus on developing practical tools and methodologies for creating ethical artificial intelligence (AI) systems, while addressing the gap between ethical principles and your application in the design process. (NA?)

  • Follow the proposed FAIRification workflow for health data, which addresses specific challenges such as ethics, legal compliance, and technical interoperability, to effectively transform raw health data into FAIR datasets suitable for sharing and reuse within the health research community. (NA?)

  • Critically evaluate existing AI ethics guidelines, identify gaps and inconsistencies, and propose improvements to ensure effective implementation and regulation of AI systems. (NA?)

  • Prioritize developing AI-driven health interventions based on local needs, health system constraints, and disease burdens in low and middle-income countries (LMICs), while ensuring ethical considerations, transparency, and rigorous evaluation through globally accessible datasets and standardized reporting guidelines. (NA?)

  • Adopt a contextual approach to studying GPT-3 and similar language models, which centers on human autonomy, a critical view of technology, and an engagement with ecologies of social harm and benefit surrounding technology design and use. (NA?)

  • Carefully examine the relationship between AI technologies and human intelligence, considering factors such as autonomy, reliability, and integration, before drawing conclusions about whether AI serves as a form of cognitive enhancement or merely as a tool for performing specific tasks. (NA?)

  • Use multiple political orientation tests to evaluate the political leanings of artificial intelligence systems like ChatGPT, as these tests can reveal underlying biases that may impact the fairness and objectivity of the system. (NA?)

Explainability And Interpretability

  • Utilize a human-in-the-loop framework in the model training process to ensure that users can observe and correct the models decision logic when confounding behaviors occur, thereby enhancing the model’s trustworthiness and performance.’ (Siyuan Yan et al. 2023)

  • Investigate and quantify the frequency and impact of disagreements between different post hoc explanation methods for machine learning models, as these disagreements can potentially lead to misleading conclusions and poor decision-making. (Krishna et al. 2022)

  • Prioritize diagnostic, debugging, adversarial, and benchmarking aspects of interpretability tools to enhance your utility for engineers in practical applications. (Räuker et al. 2022)

  • Ensure proper handling of missing data, appropriate use of imputation techniques, careful consideration of data leakage risks, and thorough assessment of statistical significance and uncertainty quantification when conducting machine learning analyses. (Pineau et al. 2020)

  • Carefully evaluate the fidelity and sensitivity of your explanatory models, considering factors like the choice of perturbations and the potential impact of noise, to ensure accurate and reliable results. (C.-K. Yeh et al. 2019)

  • Consider using the proposed FOCUS method for generating counterfactual explanations for tree-based classifiers, as it produces examples that are often significantly closer to the original instances in terms of multiple evaluation metrics, and offers flexibility depending on the chosen distance function. (Biggio and Roli 2018)

  • Combine the use of multi-armed bandits and explainable recommendations in a principled manner to achieve a balance between exploration and exploitation while improving user engagement. (McInerney et al. 2018)

  • Utilise influence functions, a classical technique from robust statistics, to effectively trace a models prediction through its learning algorithm and back to its training data. This approach helps identify the training points most responsible for a specific prediction, providing insights into model behaviour, debugging, error detection, and creation of adversarial training examples.’ (Koh and Liang 2017)

  • Ensure your attribution methods satisfy both Sensitivity and Implementation Invariance axioms, as well as completeness, to provide accurate and reliable explanations for deep neural network predictions. (Sundararajan, Taly, and Yan 2017)

  • Utilise the Contrastive Explanations Method’ (CEM) to provide comprehensive explanations for neural network classifications. This involves identifying not only the factors that must be present (‘pertinent positives’) for a particular classification, but also those that must be absent (‘pertinent negatives’). By doing so, researchers can offer more robust and nuanced explanations that better reflect the complexity of the underlying data.’ (Dhurandhar et al. 2017)

  • Employ randomization tests to ensure that saliency methods are sensitive to both the model and the data generating process, as reliance on visual assessments alone may lead to misleading conclusions. (Doshi-Velez et al. 2017)

  • Aim to develop interpretable and locally faithful explanations for machine learning models, allowing for increased trust and effective utilization of these models. (M. T. Ribeiro, Singh, and Guestrin 2016b)

  • Aim to create interpretable and locally faithful explanations for machine learning models, allowing for increased trust and effective utilization of these models. (M. T. Ribeiro, Singh, and Guestrin 2016b)

  • Use local explanation vectors to better understand the predictions of any classification method, allowing them to identify the most influential features for individual data points. (Baehrens et al. 2009)

  • Utilize Quantitative Input Influence (QII) measures to enhance the transparency of algorithmic decision-making systems, particularly those involving machine learning, by effectively capturing the degree of influence of inputs on outputs and providing a foundation for the design of transparency reports and testing tools. (NA?)

  • Use the iml package to simplify your analysis and interpretation of complex black-box machine learning models. (NA?)

  • Consider using the What-If Tool, an open-source application that allows practitioners to probe, visualize, and analyze machine learning systems, enabling them to test performance in hypothetical situations, analyze the importance of different data features, and visualize model behavior across multiple models and subsets of input data. (NA?)

  • Utilize explanation methods like Layer-wise Relevance Propagation (LRP) and SpRAy to validate and understand the behavior of your machine learning models, ensuring that they do not rely solely on test set errors or reward metrics. (NA?)

  • Prioritize understanding the practical needs and constraints of organizations when developing and deploying explainability techniques for machine learning models. (NA?)

  • Aim to generate diverse and feasible counterfactual explanations for machine learning models, balancing proximity to the original input with diversity among the counterfactuals presented, while taking into account user-defined constraints and causality. (NA?)

  • Consider the entire class of well-performing prediction models instead of focusing solely on a single model, as this allows for a more comprehensive understanding of variable importance. (NA?)

References

“145th Audio Engineering Society Convention.” 2018. https://doi.org/10.17743/aesconv.2018.978-1-942220-25-1.
“2023 IEEE/ACM 31st International Symposium on Quality of Service (IWQoS).” 2023. https://doi.org/10.1109/iwqos57198.2023.
“A 5b 800MS/s 2mW Asynchronous Binary-Search ADC in 65nm CMOS.” n.d. https://doi.org/10.1109/is.
“A Review of ChatGPT AI’s Impact on Several Business Sectors.” 2023. Zenodo, February. https://doi.org/10.5281/ZENODO.7644359.
“A Vision for the National Weather Service.” 1999, March. https://doi.org/10.17226/6434.
Abadi, Martín, and David G. Andersen. 2016. “Learning to Protect Communications with Adversarial Neural Cryptography.” arXiv. https://doi.org/10.48550/ARXIV.1610.06918.
Abadi, Martín, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, et al. 2016. “TensorFlow: A System for Large-Scale Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1605.08695.
Abbasi-Sureshjani, Samaneh, Iris Smit-Ockeloen, Erik Bekkers, Behdad Dashtbozorg, and Bart ter Haar Romeny. 2016. “Automatic Detection of Vascular Bifurcations and Crossings in Retinal Images Using Orientation Scores.” 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), April. https://doi.org/10.1109/isbi.2016.7493241.
Abbe, Emmanuel. 2017. “Community Detection and Stochastic Block Models.” arXiv. https://doi.org/10.48550/ARXIV.1703.10146.
Abdal, Rameen, Peihao Zhu, John Femiani, Niloy J. Mitra, and Peter Wonka. 2021. “CLIP2StyleGAN: Unsupervised Extraction of StyleGAN Edit Directions.” arXiv. https://doi.org/10.48550/ARXIV.2112.05219.
Abdar, Moloud, U. Rajendra Acharya, Nizal Sarrafzadegan, and Vladimir Makarenkov. 2019. “NE-Nu-SVC: A New Nested Ensemble Clinical Decision Support System for Effective Diagnosis of Coronary Artery Disease.” IEEE Access 7. https://doi.org/10.1109/access.2019.2953920.
Abd-Elaal, El-Sayed, Sithara H. P. W. Gamage, and Julie E. Mills. 2022. “Assisting Academics to Identify Computer Generated Writing.” European Journal of Engineering Education 47 (March). https://doi.org/10.1080/03043797.2022.2046709.
Abdelatty, Ahmed, Pracheta Sahoo, and Chiradeep Roy. 2018. “Structure Learning Using Forced Pruning.” arXiv. https://doi.org/10.48550/ARXIV.1812.00975.
Abdel-Hamid, Ossama, Abdel-rahman Mohamed, Hui Jiang, Li Deng, Gerald Penn, and Dong Yu. 2014. “Convolutional Neural Networks for Speech Recognition.” IEEE/ACM Transactions on Audio, Speech, and Language Processing 22 (October). https://doi.org/10.1109/taslp.2014.2339736.
Abdool, Mustafa, Malay Haldar, Prashant Ramanathan, Tyler Sax, Lanbo Zhang, Aamir Mansawala, Shulin Yang, and Thomas Legrand. 2020. “Managing Diversity in Airbnb Search.” arXiv. https://doi.org/10.48550/ARXIV.2004.02621.
Abdul-Mageed, Muhammad, AbdelRahim Elmadany, El Moatez Billah Nagoudi, Dinesh Pabbi, Kunal Verma, and Rannie Lin. 2020. “Mega-COV: A Billion-Scale Dataset of 100+ Languages for COVID-19.” arXiv. https://doi.org/10.48550/ARXIV.2005.06012.
Abel, David, Alekh Agarwal, Fernando Diaz, Akshay Krishnamurthy, and Robert E. Schapire. 2016. “Exploratory Gradient Boosting for Reinforcement Learning in Complex Domains.” arXiv. https://doi.org/10.48550/ARXIV.1603.04119.
Abel, David, Will Dabney, Anna Harutyunyan, Mark K. Ho, Michael L. Littman, Doina Precup, and Satinder Singh. 2021. “On the Expressivity of Markov Reward.” arXiv. https://doi.org/10.48550/ARXIV.2111.00876.
Aberman, Kfir, Peizhuo Li, Dani Lischinski, Olga Sorkine-Hornung, Daniel Cohen-Or, and Baoquan Chen. 2020. “Skeleton-Aware Networks for Deep Motion Retargeting.” ACM Transactions on Graphics 39 (August). https://doi.org/10.1145/3386569.3392462.
Abnar, Samira, and Willem Zuidema. 2020. “Quantifying Attention Flow in Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2005.00928.
Abraham, Alexandre, Fabian Pedregosa, Michael Eickenberg, Philippe Gervais, Andreas Muller, Jean Kossaifi, Alexandre Gramfort, Bertrand Thirion, and Gäel Varoquaux. 2014. “Machine Learning for Neuroimaging with Scikit-Learn.” arXiv. https://doi.org/10.48550/ARXIV.1412.3919.
Abreu, Rui, Peter Zoeteweij, and Arjan J. C. van Gemund. 2009. “Spectrum-Based Multiple Fault Localization.” 2009 IEEE/ACM International Conference on Automated Software Engineering, November. https://doi.org/10.1109/ase.2009.25.
Abu-Aisheh, Zeina, Romain Raveaux, Jean-Yves Ramel, and Patrick Martineau. 2015. “An Exact Graph Edit Distance Algorithm for Solving Pattern Recognition Problems.” Proceedings of the International Conference on Pattern Recognition Applications and Methods. https://doi.org/10.5220/0005209202710278.
Abu-El-Haija, Sami, Amol Kapoor, Bryan Perozzi, and Joonseok Lee. 2018. “N-GCN: Multi-Scale Graph Convolution for Semi-Supervised Node Classification.” arXiv. https://doi.org/10.48550/ARXIV.1802.08888.
Abu-El-Haija, Sami, Nisarg Kothari, Joonseok Lee, Paul Natsev, George Toderici, Balakrishnan Varadarajan, and Sudheendra Vijayanarasimhan. 2016. “YouTube-8M: A Large-Scale Video Classification Benchmark.” arXiv. https://doi.org/10.48550/ARXIV.1609.08675.
Abuzaid, Firas, Geet Sethi, Peter Bailis, and Matei Zaharia. 2019. “To Index or Not to Index: Optimizing Exact Maximum Inner Product Search.” 2019 IEEE 35th International Conference on Data Engineering (ICDE), April. https://doi.org/10.1109/icde.2019.00114.
Acharya, Anish, Rahul Goel, Angeliki Metallinou, and Inderjit Dhillon. 2018. “Online Embedding Compression for Text Classification Using Low Rank Matrix Factorization.” arXiv. https://doi.org/10.48550/ARXIV.1811.00641.
Achille, Alessandro, Tom Eccles, Loic Matthey, Christopher P. Burgess, Nick Watters, Alexander Lerchner, and Irina Higgins. 2018. “Life-Long Disentangled Representation Learning with Cross-Domain Latent Homologies.” arXiv. https://doi.org/10.48550/ARXIV.1808.06508.
Achlioptas, Panos, Olga Diamanti, Ioannis Mitliagkas, and Leonidas Guibas. 2017. “Learning Representations and Generative Models for 3D Point Clouds.” arXiv. https://doi.org/10.48550/ARXIV.1707.02392.
Ackermann, Johannes, and Minjun Li. 2022. “High-Resolution Image Editing via Multi-Stage Blended Diffusion.” arXiv. https://doi.org/10.48550/ARXIV.2210.12965.
Acosta-Mendoza, Niusvel, Andrés Gago-Alonso, Jesús Ariel Carrasco-Ochoa, José Francisco Martínez-Trinidad, and José Eladio Medina-Pagola. 2016. “Improving Graph-Based Image Classification by Using Emerging Patterns as Attributes.” Engineering Applications of Artificial Intelligence 50 (April). https://doi.org/10.1016/j.engappai.2016.01.030.
Adadi, Amina, and Mohammed Berrada. 2018. “Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI).” IEEE Access 6. https://doi.org/10.1109/access.2018.2870052.
Adams, Jadie, and Shireen Elhabian. 2023. “Fully Bayesian VIB-DeepSSM.” arXiv. https://doi.org/10.48550/ARXIV.2305.05797.
Adams, Ryan Prescott, and David J. C. MacKay. 2007. “Bayesian Online Changepoint Detection.” arXiv. https://doi.org/10.48550/ARXIV.0710.3742.
Adelson, Edward H., and James R. Bergen. 1985. “Spatiotemporal Energy Models for the Perception of Motion.” Journal of the Optical Society of America A 2 (February). https://doi.org/10.1364/josaa.2.000284.
Adhikari, Ashutosh, Xingdi Yuan, Marc-Alexandre Côté, Mikuláš Zelinka, Marc-Antoine Rondeau, Romain Laroche, Pascal Poupart, Jian Tang, Adam Trischler, and William L. Hamilton. 2020. “Learning Dynamic Belief Graphs to Generalize on Text-Based Games.” arXiv. https://doi.org/10.48550/ARXIV.2002.09127.
Adi, Yossi, Einat Kermany, Yonatan Belinkov, Ofer Lavi, and Yoav Goldberg. 2016. “Fine-Grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks.” arXiv. https://doi.org/10.48550/ARXIV.1608.04207.
Adler, Will, and Rachel Denison. 2018. “Wtadler/Confidence V1.0.0,” September. https://doi.org/10.5281/ZENODO.1422804.
“Advanced Data Mining and Applications.” 2005. Lecture Notes in Computer Science. https://doi.org/10.1007/b11111.
———. 2017. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-319-69179-4.
“Advances in Information Retrieval.” 2009. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-642-00958-7.
———. 2022. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-99739-7.
“Advances in Knowledge Discovery and Data Mining.” 2002. Lecture Notes in Computer Science. https://doi.org/10.1007/3-540-47887-6.
———. 2006. Lecture Notes in Computer Science. https://doi.org/10.1007/11731139.
Advani, Madhu S., and Andrew M. Saxe. 2017. “High-Dimensional Dynamics of Generalization Error in Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1710.03667.
Afham, Mohamed, Salman Khan, Muhammad Haris Khan, Muzammal Naseer, and Fahad Shahbaz Khan. 2021. “Rich Semantics Improve Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.2104.12709.
Afifi, Salma, Febin Sunny, Mahdi Nikdast, and Sudeep Pasricha. 2024. “Accelerating Neural Networks for Large Language Models and Graph Processing with Silicon Photonics.” arXiv. https://doi.org/10.48550/ARXIV.2401.06885.
Afshar, Parnian, Arash Mohammadi, and Konstantinos N. Plataniotis. 2018. “Brain Tumor Type Classification via Capsule Networks.” arXiv. https://doi.org/10.48550/ARXIV.1802.10200.
Agarap, Abien Fred. 2018. “Deep Learning Using Rectified Linear Units (ReLU).” arXiv. https://doi.org/10.48550/ARXIV.1803.08375.
Agarwal, Alekh, Peter L. Bartlett, Pradeep Ravikumar, and Martin J. Wainwright. 2010. “Information-Theoretic Lower Bounds on the Oracle Complexity of Stochastic Convex Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1009.0571.
Agarwal, Alekh, Sarah Bird, Markus Cozowicz, Luong Hoang, John Langford, Stephen Lee, Jiaji Li, et al. 2016. “Making Contextual Decisions with Low Technical Debt.” arXiv. https://doi.org/10.48550/ARXIV.1606.03966.
Agarwal, Alekh, Olivier Chapelle, Miroslav Dudik, and John Langford. 2011. “A Reliable Effective Terascale Linear Learning System.” arXiv. https://doi.org/10.48550/ARXIV.1110.4198.
Agarwal, Alekh, and John C. Duchi. 2011. “Distributed Delayed Stochastic Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1104.5525.
Agarwal, Aman, Ivan Zaitsev, Xuanhui Wang, Cheng Li, Marc Najork, and Thorsten Joachims. 2019. “Estimating Position Bias Without Intrusive Interventions.” Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, January. https://doi.org/10.1145/3289600.3291017.
Agarwal, Deepak, Souvik Ghosh, Kai Wei, and Siyu You. 2014. “Budget Pacing for Targeted Online Advertisements at LinkedIn.” Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2623330.2623366.
Agarwal, Naman, Zeyuan Allen-Zhu, Brian Bullins, Elad Hazan, and Tengyu Ma. 2016. “Finding Approximate Local Minima Faster Than Gradient Descent.” arXiv. https://doi.org/10.48550/ARXIV.1611.01146.
Agarwal, Prabhat, Manisha Srivastava, Vishwakarma Singh, and Charles Rosenberg. 2022. “Modeling User Behavior with Interaction Networks for Spam Detection.” Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3477495.3531875.
Aggarwal, Charu C., and Saket Sathe. 2017. “Outlier Ensembles.” https://doi.org/10.1007/978-3-319-54765-7.
Aghajanyan, Armen, Lili Yu, Alexis Conneau, Wei-Ning Hsu, Karen Hambardzumyan, Susan Zhang, Stephen Roller, Naman Goyal, Omer Levy, and Luke Zettlemoyer. 2023. “Scaling Laws for Generative Mixed-Modal Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.03728.
Aghasi, Alireza, Afshin Abdi, and Justin Romberg. 2018. “Fast Convex Pruning of Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1806.06457.
Agostinelli, Andrea, Timo I. Denk, Zalán Borsos, Jesse Engel, Mauro Verzetti, Antoine Caillon, Qingqing Huang, et al. 2023. “MusicLM: Generating Music from Text.” arXiv. https://doi.org/10.48550/ARXIV.2301.11325.
Agostinelli, Forest, Matthew Hoffman, Peter Sadowski, and Pierre Baldi. 2014. “Learning Activation Functions to Improve Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1412.6830.
Agrawal, Aishwarya, Dhruv Batra, and Devi Parikh. 2016. “Analyzing the Behavior of Visual Question Answering Models.” arXiv. https://doi.org/10.48550/ARXIV.1606.07356.
Agrawal, Aishwarya, Ivana Kajić, Emanuele Bugliarello, Elnaz Davoodi, Anita Gergely, Phil Blunsom, and Aida Nematzadeh. 2022. “Reassessing Evaluation Practices in Visual Question Answering: A Case Study on Out-of-Distribution Generalization.” arXiv. https://doi.org/10.48550/ARXIV.2205.12191.
Agrawal, Aishwarya, Jiasen Lu, Stanislaw Antol, Margaret Mitchell, C. Lawrence Zitnick, Dhruv Batra, and Devi Parikh. 2015. “VQA: Visual Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1505.00468.
Agrawal, Pulkit, Ashvin Nair, Pieter Abbeel, Jitendra Malik, and Sergey Levine. 2016. “Learning to Poke by Poking: Experiential Learning of Intuitive Physics.” arXiv. https://doi.org/10.48550/ARXIV.1606.07419.
Agrawal, Shipra, Zizhuo Wang, and Yinyu Ye. 2009. “A Dynamic Near-Optimal Algorithm for Online Linear Programming.” arXiv. https://doi.org/10.48550/ARXIV.0911.2974.
Agustsson, Eirikur, Fabian Mentzer, Michael Tschannen, Lukas Cavigelli, Radu Timofte, Luca Benini, and Luc Van Gool. 2017. “Soft-to-Hard Vector Quantization for End-to-End Learning Compressible Representations.” arXiv. https://doi.org/10.48550/ARXIV.1704.00648.
Aha, David W., Dennis Kibler, and Marc K. Albert. 1991. Machine Learning 6. https://doi.org/10.1023/a:1022689900470.
Aher, Gati, Rosa I. Arriaga, and Adam Tauman Kalai. 2022. “Using Large Language Models to Simulate Multiple Humans and Replicate Human Subject Studies.” arXiv. https://doi.org/10.48550/ARXIV.2208.10264.
Ahn, Michael, Anthony Brohan, Noah Brown, Yevgen Chebotar, Omar Cortes, Byron David, Chelsea Finn, et al. 2022. “Do as i Can, Not as i Say: Grounding Language in Robotic Affordances.” arXiv. https://doi.org/10.48550/ARXIV.2204.01691.
Ahn, Sungjin, Heeyoul Choi, Tanel Pärnamaa, and Yoshua Bengio. 2016. “A Neural Knowledge Language Model.” arXiv. https://doi.org/10.48550/ARXIV.1608.00318.
Ahn, Sungjin, Anoop Korattikara, and Max Welling. 2012. “Bayesian Posterior Sampling via Stochastic Gradient Fisher Scoring.” arXiv. https://doi.org/10.48550/ARXIV.1206.6380.
Ahuja, Kabir, Harshita Diddee, Rishav Hada, Millicent Ochieng, Krithika Ramesh, Prachi Jain, Akshay Nambi, et al. 2023. “MEGA: Multilingual Evaluation of Generative AI.” arXiv. https://doi.org/10.48550/ARXIV.2303.12528.
Ai, Xing, Jialong Zhou, Yulin Zhu, Gaolei Li, Tomasz P. Michalak, Xiapu Luo, and Kai Zhou. 2023. “Graph Anomaly Detection at Group Level: A Topology Pattern Enhanced Unsupervised Approach.” arXiv. https://doi.org/10.48550/ARXIV.2308.01063.
Ainslie, Joshua, Santiago Ontanon, Chris Alberti, Vaclav Cvicek, Zachary Fisher, Philip Pham, Anirudh Ravula, Sumit Sanghai, Qifan Wang, and Li Yang. 2020. “ETC: Encoding Long and Structured Inputs in Transformers.” Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). https://doi.org/10.18653/v1/2020.emnlp-main.19.
Airoldi, Edoardo M, David M Blei, Stephen E Fienberg, and Eric P Xing. 2007. “Mixed Membership Stochastic Blockmodels.” arXiv. https://doi.org/10.48550/ARXIV.0705.4485.
Airoldi, Edoardo M., Xiaopei Wang, and Xiaodong Lin. 2013. “Multi-Way Blockmodels for Analyzing Coordinated High-Dimensional Responses.” The Annals of Applied Statistics 7 (December). https://doi.org/10.1214/13-aoas643.
Aiyelabowo, Olumuyiwa P., Babajide A. Ojo, Suraj A. Fadare, and Oluwafemi T. Mathew. 2018. “Development of Prompt Inferno Reporting System for Curbing the Losses of Fire Outbreak.” International Journal of Electrical and Electronics Engineering 5 (December). https://doi.org/10.14445/23488379/ijeee-v5i12p101.
Ajakan, Hana, Pascal Germain, Hugo Larochelle, François Laviolette, and Mario Marchand. 2014. “Domain-Adversarial Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1412.4446.
Ajanthan, Thalaiyasingam, Puneet K. Dokania, Richard Hartley, and Philip H. S. Torr. 2018. “Proximal Mean-Field for Neural Network Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1812.04353.
Aken, Betty van, Benjamin Winter, Alexander Löser, and Felix A. Gers. 2019. “How Does BERT Answer Questions?” Proceedings of the 28th ACM International Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3357384.3358028.
Akin, Berkin, Suyog Gupta, Yun Long, Anton Spiridonov, Zhuo Wang, Marie White, Hao Xu, Ping Zhou, and Yanqi Zhou. 2022. “Searching for Efficient Neural Architectures for on-Device ML on Edge TPUs.” arXiv. https://doi.org/10.48550/ARXIV.2204.14007.
Akinwande, Victor, Yiding Jiang, Dylan Sam, and J. Zico Kolter. 2023. “Understanding Prompt Engineering May Not Require Rethinking Generalization.” arXiv. https://doi.org/10.48550/ARXIV.2310.03957.
Akoglu, Leman, and Christos Faloutsos. 2013. “Anomaly, Event, and Fraud Detection in Large Network Datasets.” Proceedings of the Sixth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/2433396.2433496.
Akoglu, Leman, Mary McGlohon, and Christos Faloutsos. 2010. “Oddball: Spotting Anomalies in Weighted Graphs.” Advances in Knowledge Discovery and Data Mining. https://doi.org/10.1007/978-3-642-13672-6_40.
Akoglu, Leman, Hanghang Tong, and Danai Koutra. 2014. “Graph Based Anomaly Detection and Description: A Survey.” Data Mining and Knowledge Discovery 29 (July). https://doi.org/10.1007/s10618-014-0365-y.
Akrami, Farahnaz, Lingbing Guo, Wei Hu, and Chengkai Li. 2018. “Re-Evaluating Embedding-Based Knowledge Graph Completion Methods.” Proceedings of the 27th ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3269206.3269266.
Akyürek, Ekin, Dale Schuurmans, Jacob Andreas, Tengyu Ma, and Denny Zhou. 2022. “What Learning Algorithm Is in-Context Learning? Investigations with Linear Models.” arXiv. https://doi.org/10.48550/ARXIV.2211.15661.
Alabdulmohsin, Ibrahim, Vinh Q. Tran, and Mostafa Dehghani. 2024. “Fractal Patterns May Unravel the Intelligence in Next-Token Prediction.” arXiv. https://doi.org/10.48550/ARXIV.2402.01825.
Alaghi, Armin, and John P. Hayes. 2014. “Fast and Accurate Computation Using Stochastic Circuits.” Design, Automation &Amp; Test in Europe Conference &Amp; Exhibition (DATE), 2014. https://doi.org/10.7873/date.2014.089.
Alahmari, Saeed S., Dmitry B. Goldgof, Peter R. Mouton, and Lawrence O. Hall. 2020. “Challenges for the Repeatability of Deep Learning Models.” IEEE Access 8. https://doi.org/10.1109/access.2020.3039833.
Alain, Guillaume, and Yoshua Bengio. 2016. “Understanding Intermediate Layers Using Linear Classifier Probes.” arXiv. https://doi.org/10.48550/ARXIV.1610.01644.
Alain, Guillaume, Alex Lamb, Chinnadhurai Sankar, Aaron Courville, and Yoshua Bengio. 2015. “Variance Reduction in SGD by Distributed Importance Sampling.” arXiv. https://doi.org/10.48550/ARXIV.1511.06481.
Alakkari, Salaheddin, and John Dingliana. 2018. “An Acceleration Scheme for Memory Limited, Streaming PCA.” arXiv. https://doi.org/10.48550/ARXIV.1807.06530.
Alamo, Teodoro, Daniel G. Reina, Martina Mammarella, and Alberto Abella. 2020. “Open Data Resources for Fighting COVID-19.” arXiv. https://doi.org/10.48550/ARXIV.2004.06111.
Alanov, Aibek, Vadim Titov, and Dmitry Vetrov. 2022. “HyperDomainNet: Universal Domain Adaptation for Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.2210.08884.
Alayrac, Jean-Baptiste, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, et al. 2022. “Flamingo: A Visual Language Model for Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.2204.14198.
Alberini, Cristina M., Sarah A. Johnson, and Xiaojing Ye. 2013. “Memory Reconsolidation.” Memory Reconsolidation. https://doi.org/10.1016/b978-0-12-386892-3.00005-6.
Albeverio, S., and M. Röckner. 1991. “Stochastic Differential Equations in Infinite Dimensions: Solutions via Dirichlet Forms.” Probability Theory and Related Fields 89 (September). https://doi.org/10.1007/bf01198791.
Albrecht, Stefano V., and Peter Stone. 2018. “Autonomous Agents Modelling Other Agents: A Comprehensive Survey and Open Problems.” Artificial Intelligence 258 (May). https://doi.org/10.1016/j.artint.2018.01.002.
Albuquerque, Pedro H. M., Denis Ribeiro do Valle, and Daijiang Li. 2019. “Bayesian LDA for Mixed-Membership Clustering Analysis: The Rlda Package.” Knowledge-Based Systems 163 (January). https://doi.org/10.1016/j.knosys.2018.10.024.
Aldarmaki, Hanan, and Mona Diab. 2019. “Context-Aware Cross-Lingual Mapping.” Proceedings of the 2019 Conference of the North. https://doi.org/10.18653/v1/n19-1391.
Aldinucci, Marco, Salvatore Ruggieri, and Massimo Torquati. 2013. “Decision Tree Building on Multi‐core Using FastFlow.” Concurrency and Computation: Practice and Experience 26 (June). https://doi.org/10.1002/cpe.3063.
Aldous, D., and P. Diaconis. 1995. “Hammersley’s Interacting Particle Process and Longest Increasing Subsequences.” Probability Theory and Related Fields 103 (June). https://doi.org/10.1007/bf01204214.
Alemi, Alex A., Francois Chollet, Niklas Een, Geoffrey Irving, Christian Szegedy, and Josef Urban. 2016. “DeepMath - Deep Sequence Models for Premise Selection.” arXiv. https://doi.org/10.48550/ARXIV.1606.04442.
Alemi, Alexander A., Ian Fischer, Joshua V. Dillon, and Kevin Murphy. 2016. “Deep Variational Information Bottleneck.” arXiv. https://doi.org/10.48550/ARXIV.1612.00410.
Alet, Ferran, Adarsh K. Jeewajee, Maria Bauza, Alberto Rodriguez, Tomas Lozano-Perez, and Leslie Pack Kaelbling. 2019. “Graph Element Networks: Adaptive, Structured Computation and Memory.” arXiv. https://doi.org/10.48550/ARXIV.1904.09019.
Alfons, Andreas, Christophe Croux, and Sarah Gelper. 2013. “Sparse Least Trimmed Squares Regression for Analyzing High-Dimensional Large Data Sets.” The Annals of Applied Statistics 7 (March). https://doi.org/10.1214/12-aoas575.
Alghamdi, Mansoor, and William Teahan. 2018. “Printed Arabic Script Recognition: A Survey.” International Journal of Advanced Computer Science and Applications 9. https://doi.org/10.14569/ijacsa.2018.090953.
Al-Hafez, Firas, Davide Tateo, Oleg Arenz, Guoping Zhao, and Jan Peters. 2023. “LS-IQ: Implicit Reward Regularization for Inverse Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.2303.00599.
Ali, Hazrat, Shafaq Murad, and Zubair Shah. 2022. “Spot the Fake Lungs: Generating Synthetic Medical Images Using Neural Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2211.00902.
Ali, Rohaid, Oliver Y. Tang, Ian D. Connolly, Patricia L. Zadnik Sullivan, John H. Shin, Jared S. Fridley, Wael F. Asaad, et al. 2023. “Performance of ChatGPT and GPT-4 on Neurosurgery Written Board Examinations,” March. https://doi.org/10.1101/2023.03.25.23287743.
Ali, Stephen R, Thomas D Dobbs, Hayley A Hutchings, and Iain S Whitaker. 2023. “Using ChatGPT to Write Patient Clinic Letters.” The Lancet Digital Health 5 (April). https://doi.org/10.1016/s2589-7500(23)00048-1.
Alipanahi, Babak, Andrew Delong, Matthew T Weirauch, and Brendan J Frey. 2015. “Predicting the Sequence Specificities of DNA- and RNA-Binding Proteins by Deep Learning.” Nature Biotechnology 33 (July). https://doi.org/10.1038/nbt.3300.
Aljalbout, Elie, Vladimir Golkov, Yawar Siddiqui, Maximilian Strobel, and Daniel Cremers. 2018. “Clustering with Deep Learning: Taxonomy and New Methods.” arXiv. https://doi.org/10.48550/ARXIV.1801.07648.
Allal, Loubna Ben, Raymond Li, Denis Kocetkov, Chenghao Mou, Christopher Akiki, Carlos Munoz Ferrandis, Niklas Muennighoff, et al. 2023. “SantaCoder: Don’t Reach for the Stars!” arXiv. https://doi.org/10.48550/ARXIV.2301.03988.
Allamanis, Miltiadis. 2018. “The Adverse Effects of Code Duplication in Machine Learning Models of Code.” arXiv. https://doi.org/10.48550/ARXIV.1812.06469.
Allamanis, Miltiadis, Earl T. Barr, Premkumar Devanbu, and Charles Sutton. 2017. “A Survey of Machine Learning for Big Code and Naturalness.” arXiv. https://doi.org/10.48550/ARXIV.1709.06182.
———. 2018. “A Survey of Machine Learning for Big Code and Naturalness.” ACM Computing Surveys 51 (July). https://doi.org/10.1145/3212695.
Allen, Brett, Brian Curless, and Zoran Popović. 2003. “The Space of Human Body Shapes.” ACM Transactions on Graphics 22 (July). https://doi.org/10.1145/882262.882311.
Allen, Carl, Ivana Balažević, and Timothy Hospedales. 2019. “Interpreting Knowledge Graph Relation Representation from Word Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1909.11611.
Allen, Kelsey R., Evan Shelhamer, Hanul Shin, and Joshua B. Tenenbaum. 2019. “Infinite Mixture Prototypes for Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.1902.04552.
Allen, Layman E., and Charles S. Saxon. 1995. “Better Language, Better Thought, Better Communication.” Proceedings of the Fifth International Conference on Artificial Intelligence and Law - ICAIL ’95. https://doi.org/10.1145/222092.222245.
Allen-Zhu, Zeyuan, Yuanzhi Li, and Zhao Song. 2018. “A Convergence Theory for Deep Learning via over-Parameterization.” arXiv. https://doi.org/10.48550/ARXIV.1811.03962.
Alm, Cecilia Ovesdotter, Dan Roth, and Richard Sproat. 2005. “Emotions from Text.” Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing - HLT ’05. https://doi.org/10.3115/1220575.1220648.
Almahairi, Amjad, Nicolas Ballas, Tim Cooijmans, Yin Zheng, Hugo Larochelle, and Aaron Courville. 2015. “Dynamic Capacity Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.07838.
Almahairi, Amjad, Sai Rajeswar, Alessandro Sordoni, Philip Bachman, and Aaron Courville. 2018. “Augmented CycleGAN: Learning Many-to-Many Mappings from Unpaired Data.” arXiv. https://doi.org/10.48550/ARXIV.1802.10151.
Alman, Josh, and Zhao Song. 2023. “Fast Attention Requires Bounded Entries.” arXiv. https://doi.org/10.48550/ARXIV.2302.13214.
Alom, Md Zahangir, Tarek M. Taha, Christopher Yakopcic, Stefan Westberg, Paheding Sidike, Mst Shamima Nasrin, Brian C Van Esesn, Abdul A S. Awwal, and Vijayan K. Asari. 2018. “The History Began from AlexNet: A Comprehensive Survey on Deep Learning Approaches.” arXiv. https://doi.org/10.48550/ARXIV.1803.01164.
Alom, Md Zahangir, Chris Yakopcic, Tarek M. Taha, and Vijayan K. Asari. 2018. “Breast Cancer Classification from Histopathological Images with Inception Recurrent Residual Convolutional Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.1811.04241.
Alon, Noga, Shay Moran, and Amir Yehudayoff. 2015. “Sign Rank Versus VC Dimension.” arXiv. https://doi.org/10.48550/ARXIV.1503.07648.
Alquier, Pierre. 2021. “User-Friendly Introduction to PAC-Bayes Bounds.” arXiv. https://doi.org/10.48550/ARXIV.2110.11216.
Al-Rfou, Rami, Dokook Choe, Noah Constant, Mandy Guo, and Llion Jones. 2018. “Character-Level Language Modeling with Deeper Self-Attention.” arXiv. https://doi.org/10.48550/ARXIV.1808.04444.
Al-Rfou, Rami, Bryan Perozzi, and Dustin Zelle. 2019. “DDGK: Learning Graph Representations for Deep Divergence Graph Kernels.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313668.
Al-Rfou, Rami, Marc Pickett, Javier Snaider, Yun-hsuan Sung, Brian Strope, and Ray Kurzweil. 2016. “Conversational Contextual Cues: The Case of Personalization and History for Response Ranking.” arXiv. https://doi.org/10.48550/ARXIV.1606.00372.
Al-Rubaie, Mohammad, and J. Morris Chang. 2018. “Privacy Preserving Machine Learning: Threats and Solutions.” arXiv. https://doi.org/10.48550/ARXIV.1804.11238.
Alschner, Wolfgang, Manfred Elsig, and Rodrigo Polanco. 2020. “Introducing the Electronic Database of Investment Treaties (EDIT): The Genesis of a New Database and Its Use.” World Trade Review 20 (August). https://doi.org/10.1017/s147474562000035x.
Alsharif, Ouais, Tom Ouyang, Francoise Beaufays, Shumin Zhai, Thomas Breuel, and Johan Schalkwyk. 2015. “Long Short Term Memory Neural Network for Keyboard Gesture Decoding.” 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), April. https://doi.org/10.1109/icassp.2015.7178336.
Alshawabkeh, Malak, Javed A. Aslam, Jennifer G. Dy, and David Kaeli. 2012. “Feature Weighting and Selection Using Hypothesis Margin of Boosting.” 2012 IEEE 12th International Conference on Data Mining, December. https://doi.org/10.1109/icdm.2012.143.
Alshawabkeh, Malak, Byunghyun Jang, and David Kaeli. 2010. “Accelerating the Local Outlier Factor Algorithm on a GPU for Intrusion Detection Systems.” Proceedings of the 3rd Workshop on General-Purpose Computation on Graphics Processing Units, March. https://doi.org/10.1145/1735688.1735707.
Altschuler, Jason, Jonathan Weed, and Philippe Rigollet. 2017. “Near-Linear Time Approximation Algorithms for Optimal Transport via Sinkhorn Iteration.” arXiv. https://doi.org/10.48550/ARXIV.1705.09634.
Alvarez, Jose, and Lars Petersson. 2016. “DecomposeMe: Simplifying ConvNets for End-to-End Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.05426.
Alvarez-Melis, David, and Tommi S. Jaakkola. 2017. “A Causal Framework for Explaining the Predictions of Black-Box Sequence-to-Sequence Models.” arXiv. https://doi.org/10.48550/ARXIV.1707.01943.
Alzantot, Moustafa, Yash Sharma, Ahmed Elgohary, Bo-Jhang Ho, Mani Srivastava, and Kai-Wei Chang. 2018. “Generating Natural Language Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1804.07998.
Alzghoul, Ahmad, and Magnus Löfstrand. 2011. “Increasing Availability of Industrial Systems Through Data Stream Mining.” Computers &Amp; Industrial Engineering 60 (March). https://doi.org/10.1016/j.cie.2010.10.008.
Amid, Ehsan, Aristides Gionis, and Antti Ukkonen. 2016. “Semi-Supervised Kernel Metric Learning Using Relative Comparisons.” arXiv. https://doi.org/10.48550/ARXIV.1612.00086.
Aminabadi, Reza Yazdani, Samyam Rajbhandari, Minjia Zhang, Ammar Ahmad Awan, Cheng Li, Du Li, Elton Zheng, et al. 2022. “DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale.” arXiv. https://doi.org/10.48550/ARXIV.2207.00032.
Amini, Arash, Marina Paez, and Lizhen Lin. 2024. “Hierarchical Stochastic Block Model for Community Detection in Multiplex Networks.” Bayesian Analysis 19 (March). https://doi.org/10.1214/22-ba1355.
Amini, Massih Reza, Tuong Vinh Truong, and Cyril Goutte. 2008. “A Boosting Algorithm for Learning Bipartite Ranking Functions with Partially Labeled Data.” Proceedings of the 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/1390334.1390354.
Amit, Ron, and Ron Meir. 2017. “Meta-Learning by Adjusting Priors Based on Extended PAC-Bayes Theory.” arXiv. https://doi.org/10.48550/ARXIV.1711.01244.
Amodei, Dario, Rishita Anubhai, Eric Battenberg, Carl Case, Jared Casper, Bryan Catanzaro, Jingdong Chen, et al. 2015. “Deep Speech 2: End-to-End Speech Recognition in English and Mandarin.” arXiv. https://doi.org/10.48550/ARXIV.1512.02595.
Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. 2016. “Concrete Problems in AI Safety.” arXiv. https://doi.org/10.48550/ARXIV.1606.06565.
Amodio, Matthew, Swarat Chaudhuri, and Thomas W. Reps. 2017. “Neural Attribute Machines for Program Generation.” arXiv. https://doi.org/10.48550/ARXIV.1705.09231.
Amoh, Justice, and Kofi M. Odame. 2019. “An Optimized Recurrent Unit for Ultra-Low-Power Keyword Spotting.” Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 3 (June). https://doi.org/10.1145/3328907.
Amos, Brandon, Laurent Dinh, Serkan Cabi, Thomas Rothörl, Sergio Gómez Colmenarejo, Alistair Muldal, Tom Erez, Yuval Tassa, Nando de Freitas, and Misha Denil. 2018. “Learning Awareness Models.” arXiv. https://doi.org/10.48550/ARXIV.1804.06318.
Amos, Brandon, and J. Zico Kolter. 2017. “OptNet: Differentiable Optimization as a Layer in Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1703.00443.
Amos, Storkey. 2008. “When Training and Test Sets Are Different: Characterizing Learning Transfer.” Dataset Shift in Machine Learning, December. https://doi.org/10.7551/mitpress/9780262170055.003.0001.
Amrane, Said, Abdallah Zahidi, Mostafa Abouricha, Nawfel Azami, Naoual Nasser, and Mohamed Errai. 2021. “Machine Learning for Monitoring of the Solenoid Valves Coil Resistance Based on Optical Fiber Squeezer.” Journal Européen Des Systèmes Automatisés 54 (October). https://doi.org/10.18280/jesa.540511.
“Analysis-Ready Standardized TCGA Data from Broad GDAC Firehose 2016_01_28 Run.” 2016. https://doi.org/10.7908/C11G0KM9.
Anandkumar, Anima, Rong Ge, Daniel Hsu, Sham M. Kakade, and Matus Telgarsky. 2012. “Tensor Decompositions for Learning Latent Variable Models.” arXiv. https://doi.org/10.48550/ARXIV.1210.7559.
Anandkumar, Animashree, Daniel Hsu, and Sham M. Kakade. 2012. “A Method of Moments for Mixture Models and Hidden Markov Models.” arXiv. https://doi.org/10.48550/ARXIV.1203.0683.
Ananthanarayanan, Rajagopal, Venkatesh Basker, Sumit Das, Ashish Gupta, Haifeng Jiang, Tianhao Qiu, Alexey Reznichenko, Deomid Ryabkov, Manpreet Singh, and Shivakumar Venkataraman. 2013. “Photon.” Proceedings of the 2013 ACM SIGMOD International Conference on Management of Data, June. https://doi.org/10.1145/2463676.2465272.
Ancona, Marco, Cengiz Öztireli, and Markus Gross. 2020. “Shapley Value as Principled Metric for Structured Network Pruning.” arXiv. https://doi.org/10.48550/ARXIV.2006.01795.
Anderson, Brandon, Truong-Son Hy, and Risi Kondor. 2019. “Cormorant: Covariant Molecular Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1906.04015.
Anderson, Peter, Angel Chang, Devendra Singh Chaplot, Alexey Dosovitskiy, Saurabh Gupta, Vladlen Koltun, Jana Kosecka, et al. 2018. “On Evaluation of Embodied Navigation Agents.” arXiv. https://doi.org/10.48550/ARXIV.1807.06757.
Anderson, Peter, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. 2017. “Bottom-up and Top-down Attention for Image Captioning and Visual Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1707.07998.
Andreas, Jacob, Dan Klein, and Sergey Levine. 2016. “Modular Multitask Reinforcement Learning with Policy Sketches.” arXiv. https://doi.org/10.48550/ARXIV.1611.01796.
Andreas, Jacob, Marcus Rohrbach, Trevor Darrell, and Dan Klein. 2015. “Neural Module Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.02799.
———. 2016. “Learning to Compose Neural Networks for Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1601.01705.
Andreev, Pavel, Aibek Alanov, Oleg Ivanov, and Dmitry Vetrov. 2023. “HIFI++: A Unified Framework for Bandwidth Extension and Speech Enhancement.” ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), June. https://doi.org/10.1109/icassp49357.2023.10097255.
Andrews, Sheldon, and Kenny Erleben. 2021. “Contact and Friction Simulation for Computer Graphics.” ACM SIGGRAPH 2021 Courses, July. https://doi.org/10.1145/3450508.3464571.
Andrieu, Christophe, and Johannes Thoms. 2008. “A Tutorial on Adaptive MCMC.” Statistics and Computing 18 (December). https://doi.org/10.1007/s11222-008-9110-y.
Angelov, Dimo. 2020. “Top2Vec: Distributed Representations of Topics.” arXiv. https://doi.org/10.48550/ARXIV.2008.09470.
Anghel, Andreea, Nikolaos Papandreou, Thomas Parnell, Alessandro De Palma, and Haralampos Pozidis. 2018. “Benchmarking and Optimization of Gradient Boosting Decision Tree Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.1809.04559.
Angrishi, Kishore. 2017. “Turning Internet of Things(IoT) into Internet of Vulnerabilities (IoV) : IoT Botnets.” arXiv. https://doi.org/10.48550/ARXIV.1702.03681.
Angryk, Rafal, Petrus Martens, Berkay Aydin, Dustin Kempton, Sushant Mahajan, Sunitha Basodi, Azim Ahmadzadeh, et al. 2020. “SWAN-SF.” https://doi.org/10.7910/DVN/EBCFKM.
Anil, Rohan, Andrew M. Dai, Orhan Firat, Melvin Johnson, Dmitry Lepikhin, Alexandre Passos, Siamak Shakeri, et al. 2023. “PaLM 2 Technical Report.” arXiv. https://doi.org/10.48550/ARXIV.2305.10403.
Anil, Rohan, Vineet Gupta, Tomer Koren, and Yoram Singer. 2019. “Memory-Efficient Adaptive Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1901.11150.
Anselmi, Fabio, Joel Z. Leibo, Lorenzo Rosasco, Jim Mutch, Andrea Tacchetti, and Tomaso Poggio. 2016. “Unsupervised Learning of Invariant Representations.” Theoretical Computer Science 633 (June). https://doi.org/10.1016/j.tcs.2015.06.048.
Antoniou, Antreas, Harrison Edwards, and Amos Storkey. 2018. “How to Train Your MAML.” arXiv. https://doi.org/10.48550/ARXIV.1810.09502.
Antoniou, Antreas, Amos Storkey, and Harrison Edwards. 2017. “Data Augmentation Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1711.04340.
Antony, Joseph, Kevin McGuinness, Noel E O Connor, and Kieran Moran. 2016. “Quantifying Radiographic Knee Osteoarthritis Severity Using Deep Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1609.02469.
Anwar, Sajid, Kyuyeon Hwang, and Wonyong Sung. 2015. “Structured Pruning of Deep Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1512.08571.
Anwar, Sajid, and Wonyong Sung. 2016. “Compact Deep Convolutional Neural Networks with Coarse Pruning.” arXiv. https://doi.org/10.48550/ARXIV.1610.09639.
Arar, Moab, Andrey Voynov, Amir Hertz, Omri Avrahami, Shlomi Fruchter, Yael Pritch, Daniel Cohen-Or, and Ariel Shamir. 2024. “PALP: Prompt Aligned Personalization of Text-to-Image Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.06105.
Arawjo, Ian, Chelse Swoopes, Priyan Vaithilingam, Martin Wattenberg, and Elena Glassman. 2023. “ChainForge: A Visual Toolkit for Prompt Engineering and LLM Hypothesis Testing.” arXiv. https://doi.org/10.48550/ARXIV.2309.09128.
Arbabshirani, Mohammad R., Brandon K. Fornwalt, Gino J. Mongelluzzo, Jonathan D. Suever, Brandon D. Geise, Aalpen A. Patel, and Gregory J. Moore. 2018. “Advanced Machine Learning in Action: Identification of Intracranial Hemorrhage on Computed Tomography Scans of the Head with Clinical Workflow Integration.” Npj Digital Medicine 1 (April). https://doi.org/10.1038/s41746-017-0015-z.
Archer, Evan, Il Memming Park, Lars Buesing, John Cunningham, and Liam Paninski. 2015. “Black Box Variational Inference for State Space Models.” arXiv. https://doi.org/10.48550/ARXIV.1511.07367.
Ardizzone, Lynton, Carsten Lüth, Jakob Kruse, Carsten Rother, and Ullrich Köthe. 2019. “Guided Image Generation with Conditional Invertible Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1907.02392.
Arevalo, John, Thamar Solorio, Manuel Montes-y-Gómez, and Fabio A. González. 2017. “Gated Multimodal Units for Information Fusion.” arXiv. https://doi.org/10.48550/ARXIV.1702.01992.
Arguin, Louis-Pierre, Anton Bovier, and Nicola Kistler. 2012. “The Extremal Process of Branching Brownian Motion.” Probability Theory and Related Fields 157 (November). https://doi.org/10.1007/s00440-012-0464-x.
Arik, Sercan O., Jitong Chen, Kainan Peng, Wei Ping, and Yanqi Zhou. 2018. “Neural Voice Cloning with a Few Samples.” arXiv. https://doi.org/10.48550/ARXIV.1802.06006.
Arik, Sercan, Gregory Diamos, Andrew Gibiansky, John Miller, Kainan Peng, Wei Ping, Jonathan Raiman, and Yanqi Zhou. 2017. “Deep Voice 2: Multi-Speaker Neural Text-to-Speech.” arXiv. https://doi.org/10.48550/ARXIV.1705.08947.
Arivazhagan, Manoj Ghuhan, Vinay Aggarwal, Aaditya Kumar Singh, and Sunav Choudhary. 2019. “Federated Learning with Personalization Layers.” arXiv. https://doi.org/10.48550/ARXIV.1912.00818.
Arjovsky, Martin, and Léon Bottou. 2017. “Towards Principled Methods for Training Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1701.04862.
Arjovsky, Martin, Soumith Chintala, and Léon Bottou. 2017. “Wasserstein GAN.” arXiv. https://doi.org/10.48550/ARXIV.1701.07875.
Arora, Daman, and Himanshu Gaurav Singh. 2023. “Have LLMs Advanced Enough? A Challenging Problem Solving Benchmark for Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2305.15074.
Arora, Raman, Ofer Dekel, and Ambuj Tewari. 2012. “Online Bandit Learning Against an Adaptive Adversary: From Regret to Policy Regret.” arXiv. https://doi.org/10.48550/ARXIV.1206.6400.
Arora, Sanjeev, Rong Ge, Yingyu Liang, Tengyu Ma, and Yi Zhang. 2017. “Generalization and Equilibrium in Generative Adversarial Nets (GANs).” arXiv. https://doi.org/10.48550/ARXIV.1703.00573.
Arora, Sanjeev, Rong Ge, Behnam Neyshabur, and Yi Zhang. 2018. “Stronger Generalization Bounds for Deep Nets via a Compression Approach.” arXiv. https://doi.org/10.48550/ARXIV.1802.05296.
Arora, Siddhant. 2020. “A Survey on Graph Neural Networks for Knowledge Graph Completion,” July. http://arxiv.org/abs/2007.12374v1.
Arous, G. Ben, A. Dembo, and A. Guionnet. 2001. “Aging of Spherical Spin Glasses.” Probability Theory and Related Fields 120 (May). https://doi.org/10.1007/pl00008774.
Arous, G. Ben, and A. Guionnet. 1997. “Large Deviations for Wigner’s Law and Voiculescu’s Non-Commutative Entropy.” Probability Theory and Related Fields 108 (August). https://doi.org/10.1007/s004400050119.
Arpit, Devansh, Stanisław Jastrzębski, Nicolas Ballas, David Krueger, Emmanuel Bengio, Maxinder S. Kanwal, Tegan Maharaj, et al. 2017. “A Closer Look at Memorization in Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1706.05394.
Artetxe, Mikel, Gorka Labaka, and Eneko Agirre. 2018. “A Robust Self-Learning Method for Fully Unsupervised Cross-Lingual Mappings of Word Embeddings.” Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/p18-1073.
Artetxe, Mikel, Gorka Labaka, Eneko Agirre, and Kyunghyun Cho. 2017. “Unsupervised Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1710.11041.
Artetxe, Mikel, and Holger Schwenk. 2019. “Margin-Based Parallel Corpus Mining with Multilingual Sentence Embeddings.” Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/p19-1309.
“Artificial Neural Networks and Machine Learning – ICANN 2018.” 2018. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-01418-6.
Aryafar, Kamelia, Devin Guillory, and Liangjie Hong. 2017. “An Ensemble-Based Approach to Click-Through Rate Prediction for Promoted Listings at Etsy.” Proceedings of the ADKDD’17, August. https://doi.org/10.1145/3124749.3124758.
Ashbrook, Daniel, Patrick Baudisch, and Sean White. 2011. “Nenya.” Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, May. https://doi.org/10.1145/1978942.1979238.
Ashok, Anubhav, Nicholas Rhinehart, Fares Beainy, and Kris M. Kitani. 2017. “N2N Learning: Network to Network Compression via Policy Gradient Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1709.06030.
Asif, Umar, Jianbin Tang, and Stefan Harrer. 2019. “Ensemble Knowledge Distillation for Learning Improved and Efficient Networks.” arXiv. https://doi.org/10.48550/ARXIV.1909.08097.
Asim, Muhammad, Fahad Shamshad, and Ali Ahmed. 2018. “Blind Image Deconvolution Using Deep Generative Priors.” arXiv. https://doi.org/10.48550/ARXIV.1802.04073.
Askari, Armin, Geoffrey Negiar, Rajiv Sambharya, and Laurent El Ghaoui. 2018. “Lifted Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1805.01532.
Asparouhov, Tihomir, and Bengt Muthén. 2014. “Auxiliary Variables in Mixture Modeling: Three-Step Approaches Using m<i>plus</i>.” Structural Equation Modeling: A Multidisciplinary Journal 21 (June). https://doi.org/10.1080/10705511.2014.915181.
Asri, Layla El, Jing He, and Kaheer Suleman. 2016. “A Sequence-to-Sequence Model for User Simulation in Spoken Dialogue Systems.” arXiv. https://doi.org/10.48550/ARXIV.1607.00070.
Assran, Mahmoud, Nicolas Ballas, Lluis Castrejon, and Michael Rabbat. 2020. “Supervision Accelerates Pre-Training in Contrastive Semi-Supervised Learning of Visual Representations.” arXiv. https://doi.org/10.48550/ARXIV.2006.10803.
Assran, Mahmoud, Mathilde Caron, Ishan Misra, Piotr Bojanowski, Florian Bordes, Pascal Vincent, Armand Joulin, Michael Rabbat, and Nicolas Ballas. 2022. “Masked Siamese Networks for Label-Efficient Learning.” arXiv. https://doi.org/10.48550/ARXIV.2204.07141.
Astrid, Marcella, and Seung-Ik Lee. 2017. “CP-Decomposition with Tensor Power Method for Convolutional Neural Networks Compression.” arXiv. https://doi.org/10.48550/ARXIV.1701.07148.
Atefeh, Farzindar, and Wael Khreich. 2013. “A Survey of Techniques for Event Detection in Twitter.” Computational Intelligence 31 (September). https://doi.org/10.1111/coin.12017.
Ateniese, Giuseppe, Giovanni Felici, Luigi V. Mancini, Angelo Spognardi, Antonio Villani, and Domenico Vitali. 2013. “Hacking Smart Machines with Smarter Ones: How to Extract Meaningful Data from Machine Learning Classifiers.” arXiv. https://doi.org/10.48550/ARXIV.1306.4447.
Athalye, Anish, Logan Engstrom, Andrew Ilyas, and Kevin Kwok. 2017. “Synthesizing Robust Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1707.07397.
Athey, Susan, Julie Tibshirani, and Stefan Wager. 2019. “Generalized Random Forests.” The Annals of Statistics 47 (April). https://doi.org/10.1214/18-aos1709.
Atienza, Rowel. 2021. “Vision Transformer for Fast and Efficient Scene Text Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2105.08582.
Atwood, James, and Don Towsley. 2015. “Diffusion-Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.02136.
Atzori, Luigi, Antonio Iera, and Giacomo Morabito. 2010. “The Internet of Things: A Survey.” Computer Networks 54 (October). https://doi.org/10.1016/j.comnet.2010.05.010.
Au, Timothy C. 2017. “Random Forests, Decision Trees, and Categorical Predictors: The "Absent Levels" Problem.” arXiv. https://doi.org/10.48550/ARXIV.1706.03492.
Aubry, Mathieu, Sylvain Paris, Samuel W. Hasinoff, Jan Kautz, and Frédo Durand. 2014. “Fast Local Laplacian Filters.” ACM Transactions on Graphics 33 (September). https://doi.org/10.1145/2629645.
Audhkhasi, Kartik, Bhuvana Ramabhadran, George Saon, Michael Picheny, and David Nahamoo. 2017. “Direct Acoustics-to-Word Models for English Conversational Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1703.07754.
“Augmented Cognition.” 2021. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-78114-9.
Ausat, Abu Muna Almaududi. 2022. “Positive Impact of the Covid-19 Pandemic on the World of Education.” Jurnal Pendidikan 23 (September). https://doi.org/10.33830/jp.v23i2.3048.2022.
Auvolat, Alex, Sarath Chandar, Pascal Vincent, Hugo Larochelle, and Yoshua Bengio. 2015. “Clustering Is Efficient for Approximate Maximum Inner Product Search.” arXiv. https://doi.org/10.48550/ARXIV.1507.05910.
Avalos-Pacheco, Alejandra, David Rossell, and Richard S. Savage. 2022. “Heterogeneous Large Datasets Integration Using Bayesian Factor Regression.” Bayesian Analysis 17 (March). https://doi.org/10.1214/20-ba1240.
Avnimelech, Ran, and Nathan Intrator. 1999. “Boosted Mixture of Experts: An Ensemble Learning Scheme.” Neural Computation 11 (February). https://doi.org/10.1162/089976699300016737.
Aybat, Necdet Serhat, Zi Wang, Tianyi Lin, and Shiqian Ma. 2015. “Distributed Linearized Alternating Direction Method of Multipliers for Composite Convex Consensus Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1512.08122.
Aydın, Burcu, Gábor Pataki, Haonan Wang, Elizabeth Bullitt, and J. S. Marron. 2009. “A Principal Component Analysis for Trees.” The Annals of Applied Statistics 3 (December). https://doi.org/10.1214/09-aoas263.
Azadi, Samaneh, Jiashi Feng, Stefanie Jegelka, and Trevor Darrell. 2015. “Auxiliary Image Regularization for Deep CNNs with Noisy Labels.” arXiv. https://doi.org/10.48550/ARXIV.1511.07069.
Azar, Edward E. 1984. “Conflict and Peace Data Bank (COPDAB), 1948-1978.” https://doi.org/10.3886/ICPSR07767.V4.
Ba, Jimmy Lei, Jamie Ryan Kiros, and Geoffrey E. Hinton. 2016. “Layer Normalization.” arXiv. https://doi.org/10.48550/ARXIV.1607.06450.
Ba, Jimmy, Volodymyr Mnih, and Koray Kavukcuoglu. 2014. “Multiple Object Recognition with Visual Attention.” arXiv. https://doi.org/10.48550/ARXIV.1412.7755.
Ba, Shan, and V. Roshan Joseph. 2012. “Composite Gaussian Process Models for Emulating Expensive Functions.” The Annals of Applied Statistics 6 (December). https://doi.org/10.1214/12-aoas570.
Babenko, Artem, and Victor Lempitsky. 2014. “Improving Bilayer Product Quantization for Billion-Scale Approximate Nearest Neighbors in High Dimensions.” arXiv. https://doi.org/10.48550/ARXIV.1404.1831.
Babin, D., I. Mazurenko, D. Parkhomenko, and A. Voloshko. 2018. “CNN Inference Acceleration Using Dictionary of Centroids.” arXiv. https://doi.org/10.48550/ARXIV.1810.08612.
Bach, Francis. 2008. “Bolasso: Model Consistent Lasso Estimation Through the Bootstrap.” arXiv. https://doi.org/10.48550/ARXIV.0804.1302.
———. 2010. “Convex Analysis and Optimization with Submodular Functions: A Tutorial.” arXiv. https://doi.org/10.48550/ARXIV.1010.4207.
———. 2011. “Learning with Submodular Functions: A Convex Optimization Perspective.” arXiv. https://doi.org/10.48550/ARXIV.1111.6453.
Bach, Sebastian, Alexander Binder, Grégoire Montavon, Frederick Klauschen, Klaus-Robert Müller, and Wojciech Samek. 2015. “On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.” PLOS ONE 10 (July). https://doi.org/10.1371/journal.pone.0130140.
Bach, Stephen H., Bryan He, Alexander Ratner, and Christopher Ré. 2017. “Learning the Structure of Generative Models Without Labeled Data.” arXiv. https://doi.org/10.48550/ARXIV.1703.00854.
Bach, Stephen H., Victor Sanh, Zheng-Xin Yong, Albert Webson, Colin Raffel, Nihal V. Nayak, Abheesht Sharma, et al. 2022. “PromptSource: An Integrated Development Environment and Repository for Natural Language Prompts.” arXiv. https://doi.org/10.48550/ARXIV.2202.01279.
Bachmann, Gregor, Gary Bécigneul, and Octavian-Eugen Ganea. 2019. “Constant Curvature Graph Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1911.05076.
Bader, Bret, Richard A. Harshman, and Tamara G. Kolda. 2007. “Temporal Analysis of Semantic Graphs Using ASALSAN.” Seventh IEEE International Conference on Data Mining (ICDM 2007), October. https://doi.org/10.1109/icdm.2007.54.
Badillo, Solveig, Balazs Banfai, Fabian Birzele, Iakov I. Davydov, Lucy Hutchinson, Tony Kam‐Thong, Juliane Siebourg‐Polster, Bernhard Steiert, and Jitao David Zhang. 2020. “An Introduction to Machine Learning.” Clinical Pharmacology &Amp; Therapeutics 107 (March). https://doi.org/10.1002/cpt.1796.
Badlani, Rohan, Adrian Łancucki, Kevin J. Shih, Rafael Valle, Wei Ping, and Bryan Catanzaro. 2021. “One TTS Alignment to Rule Them All.” arXiv. https://doi.org/10.48550/ARXIV.2108.10447.
Badrinarayanan, Vijay, Alex Kendall, and Roberto Cipolla. 2015. “SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1511.00561.
Baehrens, David, Timon Schroeter, Stefan Harmeling, Motoaki Kawanabe, Katja Hansen, and Klaus-Robert Mueller. 2009. “How to Explain Individual Classification Decisions.” arXiv. https://doi.org/10.48550/ARXIV.0912.1128.
Baevski, Alexei, Michael Auli, and Abdelrahman Mohamed. 2019. “Effectiveness of Self-Supervised Pre-Training for Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1911.03912.
Baevski, Alexei, Steffen Schneider, and Michael Auli. 2019. “Vq-Wav2vec: Self-Supervised Learning of Discrete Speech Representations.” arXiv. https://doi.org/10.48550/ARXIV.1910.05453.
Baevski, Alexei, Henry Zhou, Abdelrahman Mohamed, and Michael Auli. 2020. “Wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations.” arXiv. https://doi.org/10.48550/ARXIV.2006.11477.
Bagherinezhad, Hessam, Maxwell Horton, Mohammad Rastegari, and Ali Farhadi. 2018. “Label Refinery: Improving ImageNet Classification Through Label Progression.” arXiv. https://doi.org/10.48550/ARXIV.1805.02641.
Bagherinezhad, Hessam, Mohammad Rastegari, and Ali Farhadi. 2016. “LCNN: Lookup-Based Convolutional Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.1611.06473.
Baglio, Julien, Le Duc Ninh, and Marcus M. Weber. 2016. “Erratum: Massive Gauge Boson Pair Production at the LHC: A Next-to-Leading Order Story [Phys. Rev. D <b>88</b> , 113005 (2013)].” Physical Review D 94 (November). https://doi.org/10.1103/physrevd.94.099902.
Bagnall, Anthony, Aaron Bostrom, James Large, and Jason Lines. 2016. “The Great Time Series Classification Bake Off: An Experimental Evaluation of Recently Proposed Algorithms. Extended Version.” arXiv. https://doi.org/10.48550/ARXIV.1602.01711.
Bagnall, Anthony, Hoang Anh Dau, Jason Lines, Michael Flynn, James Large, Aaron Bostrom, Paul Southam, and Eamonn Keogh. 2018. “The UEA Multivariate Time Series Classification Archive, 2018.” arXiv. https://doi.org/10.48550/ARXIV.1811.00075.
Bahack, Lear. 2013. “Theoretical Bitcoin Attacks with Less Than Half of the Computational Power (Draft).” arXiv. https://doi.org/10.48550/ARXIV.1312.7013.
Bahdanau, Dzmitry, Philemon Brakel, Kelvin Xu, Anirudh Goyal, Ryan Lowe, Joelle Pineau, Aaron Courville, and Yoshua Bengio. 2016. “An Actor-Critic Algorithm for Sequence Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1607.07086.
Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2014. “Neural Machine Translation by Jointly Learning to Align and Translate,” September. http://arxiv.org/abs/1409.0473v7.
Bahdanau, Dzmitry, Jan Chorowski, Dmitriy Serdyuk, Philemon Brakel, and Yoshua Bengio. 2015. “End-to-End Attention-Based Large Vocabulary Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1508.04395.
Bahdanau, Dzmitry, Dmitriy Serdyuk, Philémon Brakel, Nan Rosemary Ke, Jan Chorowski, Aaron Courville, and Yoshua Bengio. 2015. “Task Loss Estimation for Sequence Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1511.06456.
Bahng, Hyojin, Ali Jahanian, Swami Sankaranarayanan, and Phillip Isola. 2022. “Exploring Visual Prompts for Adapting Large-Scale Models.” arXiv. https://doi.org/10.48550/ARXIV.2203.17274.
Bai, Jinze, Shuai Bai, Yunfei Chu, Zeyu Cui, Kai Dang, Xiaodong Deng, Yang Fan, et al. 2023. “Qwen Technical Report.” arXiv. https://doi.org/10.48550/ARXIV.2309.16609.
Bai, Shuanghao, Min Zhang, Wanqi Zhou, Siteng Huang, Zhirong Luan, Donglin Wang, and Badong Chen. 2023. “Prompt-Based Distribution Alignment for Unsupervised Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.2312.09553.
Bai, Xiaole, Santosh Kumar, Dong Xuan, Ziqiu Yun, and Ten H. Lai. 2006. “Deploying Wireless Sensors to Achieve Both Coverage and Connectivity.” Proceedings of the 7th ACM International Symposium on Mobile Ad Hoc Networking and Computing, May. https://doi.org/10.1145/1132905.1132921.
Bai, Yingbin, Erkun Yang, Bo Han, Yanhua Yang, Jiatong Li, Yinian Mao, Gang Niu, and Tongliang Liu. 2021. “Understanding and Improving Early Stopping for Learning with Noisy Labels.” arXiv. https://doi.org/10.48550/ARXIV.2106.15853.
Bain, Max, Arsha Nagrani, Gül Varol, and Andrew Zisserman. 2022. “A CLIP-Hitchhiker’s Guide to Long Video Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.2205.08508.
Baio, Gianluca, and Marta Blangiardo. 2010. “Bayesian Hierarchical Model for the Prediction of Football Results.” Journal of Applied Statistics 37 (January). https://doi.org/10.1080/02664760802684177.
Bakas, Spyridon, Mauricio Reyes, Andras Jakab, Stefan Bauer, Markus Rempfler, Alessandro Crimi, Russell Takeshi Shinohara, et al. 2018. “Identifying the Best Machine Learning Algorithms for Brain Tumor Segmentation, Progression Assessment, and Overall Survival Prediction in the BRATS Challenge.” arXiv. https://doi.org/10.48550/ARXIV.1811.02629.
Baker, Bowen, Otkrist Gupta, Nikhil Naik, and Ramesh Raskar. 2016. “Designing Neural Network Architectures Using Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.02167.
Baker, Bowen, Otkrist Gupta, Ramesh Raskar, and Nikhil Naik. 2017. “Accelerating Neural Architecture Search Using Performance Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1705.10823.
Baker, Collin F., Charles J. Fillmore, and John B. Lowe. 1998. “The Berkeley FrameNet Project.” Proceedings of the 36th Annual Meeting on Association for Computational Linguistics -. https://doi.org/10.3115/980845.980860.
Baker, Matthew E., and Ryan S. King. 2010. “A New Method for Detecting and Interpreting Biodiversity and Ecological Community Thresholds.” Methods in Ecology and Evolution 1 (February). https://doi.org/10.1111/j.2041-210x.2009.00007.x.
Baker, Nathan, Frank Alexander, Timo Bremer, Aric Hagberg, Yannis Kevrekidis, Habib Najm, Manish Parashar, et al. 2018. “Brochure on Basic Research Needs for Scientific Machine Learning: Core Technologies for Artificial Intelligence,” December. https://doi.org/10.2172/1484362.
Balaji, Yogesh, Seungjun Nah, Xun Huang, Arash Vahdat, Jiaming Song, Qinsheng Zhang, Karsten Kreis, et al. 2022. “eDiff-i: Text-to-Image Diffusion Models with an Ensemble of Expert Denoisers.” arXiv. https://doi.org/10.48550/ARXIV.2211.01324.
Balazevic, Ivana, Carl Allen, and Timothy Hospedales. 2019. “TuckER: Tensor Factorization for Knowledge Graph Completion.” Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). https://doi.org/10.18653/v1/d19-1522.
Baldassarre, Federico, and Hossein Azizpour. 2019. “Explainability Techniques for Graph Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1905.13686.
Baldassi, Carlo, Alessandro Ingrosso, Carlo Lucibello, Luca Saglietti, and Riccardo Zecchina. 2015. “Subdominant Dense Clusters Allow for Simple Learning and High Computational Performance in Neural Networks with Discrete Synapses.” Physical Review Letters 115 (September). https://doi.org/10.1103/physrevlett.115.128101.
Baldassini, Folco Bertini, Huy H. Nguyen, Ching-Chung Chang, and Isao Echizen. 2024. “Cross-Attention Watermarking of Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.06829.
Balduzzi, David, and Muhammad Ghifary. 2015. “Compatible Value Gradients for Reinforcement Learning of Continuous Deep Policies.” arXiv. https://doi.org/10.48550/ARXIV.1509.03005.
Balestriero, Randall, Mark Ibrahim, Vlad Sobal, Ari Morcos, Shashank Shekhar, Tom Goldstein, Florian Bordes, et al. 2023. “A Cookbook of Self-Supervised Learning.” arXiv. https://doi.org/10.48550/ARXIV.2304.12210.
Ball, Richard D., Valerio Bertone, Stefano Carrazza, Christopher S. Deans, Luigi Del Debbio, Stefano Forte, Alberto Guffanti, et al. 2013. “Parton Distributions with LHC Data.” Nuclear Physics B 867 (February). https://doi.org/10.1016/j.nuclphysb.2012.10.003.
Ballé, Johannes, Valero Laparra, and Eero P. Simoncelli. 2016. “End-to-End Optimized Image Compression.” arXiv. https://doi.org/10.48550/ARXIV.1611.01704.
Ballesteros, Miguel, Chris Dyer, and Noah A. Smith. 2015. “Improved Transition-Based Parsing by Modeling Characters Instead of Words with LSTMs.” arXiv. https://doi.org/10.48550/ARXIV.1508.00657.
Bally, V., and D. Talay. 1996. “The Law of the Euler Scheme for Stochastic Differential Equations.” Probability Theory and Related Fields 104 (March). https://doi.org/10.1007/bf01303802.
Balog, Matej, Alexander L. Gaunt, Marc Brockschmidt, Sebastian Nowozin, and Daniel Tarlow. 2016. “DeepCoder: Learning to Write Programs.” arXiv. https://doi.org/10.48550/ARXIV.1611.01989.
Baluja, Shumeet, and Ian Fischer. 2017. “Adversarial Transformation Networks: Learning to Generate Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1703.09387.
Banbury, Colby, Vijay Janapa Reddi, Peter Torelli, Jeremy Holleman, Nat Jeffries, Csaba Kiraly, Pietro Montino, et al. 2021. “MLPerf Tiny Benchmark.” arXiv. https://doi.org/10.48550/ARXIV.2106.07597.
Banderier, C., and P. Nicodème. 2010. “Bounded Discrete Walks.” Discrete Mathematics &Amp; Theoretical Computer Science DMTCS Proceedings vol. AM,... (January). https://doi.org/10.46298/dmtcs.2792.
Bang, Seungbae, Maria Korosteleva, and Sung‐Hee Lee. 2021. “Estimating Garment Patterns from Static Scan Data.” Computer Graphics Forum 40 (May). https://doi.org/10.1111/cgf.14272.
Bang, Yejin, Samuel Cahyawijaya, Nayeon Lee, Wenliang Dai, Dan Su, Bryan Wilie, Holy Lovenia, et al. 2023. “A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity.” arXiv. https://doi.org/10.48550/ARXIV.2302.04023.
Banner, Ron, Yury Nahshan, Elad Hoffer, and Daniel Soudry. 2018. “Post-Training 4-Bit Quantization of Convolution Networks for Rapid-Deployment.” arXiv. https://doi.org/10.48550/ARXIV.1810.05723.
Bansal, Nikhil, Avrim Blum, and Shuchi Chawla. 2004. “Correlation Clustering.” Machine Learning 56 (July). https://doi.org/10.1023/b:mach.0000033116.57574.95.
Bansal, Parikshit, Prathamesh Deshpande, and Sunita Sarawagi. 2021. “Missing Value Imputation on Multidimensional Time Series.” Proceedings of the VLDB Endowment 14 (July). https://doi.org/10.14778/3476249.3476300.
Bansal, Trapit, Jakub Pachocki, Szymon Sidor, Ilya Sutskever, and Igor Mordatch. 2017. “Emergent Complexity via Multi-Agent Competition.” arXiv. https://doi.org/10.48550/ARXIV.1710.03748.
Bao, Hangbo, Li Dong, Songhao Piao, and Furu Wei. 2021. “BEiT: BERT Pre-Training of Image Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2106.08254.
Bao, Siqi, Huang He, Fan Wang, Hua Wu, Haifeng Wang, Wenquan Wu, Zhen Guo, Zhibin Liu, and Xinchao Xu. 2020. “PLATO-2: Towards Building an Open-Domain Chatbot via Curriculum Learning.” arXiv. https://doi.org/10.48550/ARXIV.2006.16779.
Bapna, Ankur, Naveen Arivazhagan, and Orhan Firat. 2019. “Simple, Scalable Adaptation for Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1909.08478.
Baqapuri, Afroze Ibrahim, and Ilya Trofimov. 2014. “Using Neural Networks for Click Prediction of Sponsored Search.” arXiv. https://doi.org/10.48550/ARXIV.1412.6601.
Bar, Noga, Tomer Koren, and Raja Giryes. 2021. “Multiplicative Reweighting for Robust Neural Network Optimization.” arXiv. https://doi.org/10.48550/ARXIV.2102.12192.
Barabási, Albert-László, and Réka Albert. 1999. “Emergence of Scaling in Random Networks.” Science 286 (October). https://doi.org/10.1126/science.286.5439.509.
Bardes, Adrien, Jean Ponce, and Yann LeCun. 2021. “VICReg: Variance-Invariance-Covariance Regularization for Self-Supervised Learning,” May. http://arxiv.org/abs/2105.04906v3.
———. 2022. “VICRegL: Self-Supervised Learning of Local Visual Features.” arXiv. https://doi.org/10.48550/ARXIV.2210.01571.
Barkan, Oren, and Noam Koenigstein. 2016. “Item2Vec: Neural Item Embedding for Collaborative Filtering.” arXiv. https://doi.org/10.48550/ARXIV.1603.04259.
Barlow, H. B. 1989. “Unsupervised Learning.” Neural Computation 1 (September). https://doi.org/10.1162/neco.1989.1.3.295.
Barnes, Richard, Senaka Buthpitiya, James Cook, Alex Fabrikant, Andrew Tomkins, and Fangzhou Xu. 2020. “BusTr.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403376.
Barocas, Solon, Anhong Guo, Ece Kamar, Jacquelyn Krones, Meredith Ringel Morris, Jennifer Wortman Vaughan, W. Duncan Wadsworth, and Hanna Wallach. 2021. “Designing Disaggregated Evaluations of AI Systems: Choices, Considerations, and Tradeoffs.” Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, July. https://doi.org/10.1145/3461702.3462610.
Bar-On, Yinon M., Avi I. Flamholz, Rob Phillips, and Ron Milo. 2020. “SARS-CoV-2 (COVID-19) by the Numbers.” arXiv. https://doi.org/10.48550/ARXIV.2003.12886.
Barrau, Axel, and Silvere Bonnabel. 2015. “An EKF-SLAM Algorithm with Consistency Properties.” arXiv. https://doi.org/10.48550/ARXIV.1510.06263.
Barreno, Marco, Blaine Nelson, Russell Sears, Anthony D. Joseph, and J. D. Tygar. 2006. “Can Machine Learning Be Secure?” Proceedings of the 2006 ACM Symposium on Information, Computer and Communications Security, March. https://doi.org/10.1145/1128817.1128824.
Barrett, Clark, Brad Boyd, Elie Bursztein, Nicholas Carlini, Brad Chen, Jihye Choi, Amrita Roy Chowdhury, et al. 2023. “Identifying and Mitigating the Security Risks of Generative AI.” Foundations and Trends® in Privacy and Security 6. https://doi.org/10.1561/3300000041.
Barron, Andrew, Lucien Birgé, and Pascal Massart. 1999. “Risk Bounds for Model Selection via Penalization.” Probability Theory and Related Fields 113 (February). https://doi.org/10.1007/s004400050210.
Bartlett, Peter L., Paul Fischer, and Klaus-Uwe Höffgen. 2002. “Exploiting Random Walks for Learning.” Information and Computation 176 (August). https://doi.org/10.1006/inco.2002.3083.
Bartlett, Peter, Yoav Freund, Wee Sun Lee, and Robert E. Schapire. 1998. “Boosting the Margin: A New Explanation for the Effectiveness of Voting Methods.” The Annals of Statistics 26 (October). https://doi.org/10.1214/aos/1024691352.
Baryshnikov, Yu. 2001. “GUEs and Queues.” Probability Theory and Related Fields 119 (February). https://doi.org/10.1007/pl00008760.
Basir, Shamsulhaq, and Inanc Senocak. 2022. “Physics and Equality Constrained Artificial Neural Networks: Application to Forward and Inverse Problems with Multi-Fidelity Data Fusion.” Journal of Computational Physics 463 (August). https://doi.org/10.1016/j.jcp.2022.111301.
Baskin, Chaim, Natan Liss, Eli Schwartz, Evgenii Zheltonozhskii, Raja Giryes, Alex M. Bronstein, and Avi Mendelson. 2019. “UNIQ.” ACM Transactions on Computer Systems 37 (November). https://doi.org/10.1145/3444943.
Bastani, Osbert, Carolyn Kim, and Hamsa Bastani. 2017. “Interpretability via Model Extraction.” arXiv. https://doi.org/10.48550/ARXIV.1706.09773.
Bastien, Frédéric, Pascal Lamblin, Razvan Pascanu, James Bergstra, Ian Goodfellow, Arnaud Bergeron, Nicolas Bouchard, David Warde-Farley, and Yoshua Bengio. 2012. “Theano: New Features and Speed Improvements.” arXiv. https://doi.org/10.48550/ARXIV.1211.5590.
Bastings, Jasmijn, Ivan Titov, Wilker Aziz, Diego Marcheggiani, and Khalil Sima’an. 2017. “Graph Convolutional Encoders for Syntax-Aware Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1704.04675.
Battaglia, Peter W., Jessica B. Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zambaldi, Mateusz Malinowski, Andrea Tacchetti, et al. 2018. “Relational Inductive Biases, Deep Learning, and Graph Networks,” June. http://arxiv.org/abs/1806.01261v3.
Battaglia, Peter W., Razvan Pascanu, Matthew Lai, Danilo Rezende, and Koray Kavukcuoglu. 2016. “Interaction Networks for Learning about Objects, Relations and Physics.” arXiv. https://doi.org/10.48550/ARXIV.1612.00222.
Battenberg, Eric, Rewon Child, Adam Coates, Christopher Fougner, Yashesh Gaur, Jiaji Huang, Heewoo Jun, et al. 2017. “Reducing Bias in Production Speech Models.” arXiv. https://doi.org/10.48550/ARXIV.1705.04400.
Baturo, Alexander, Niheer Dasandi, and Slava J. Mikhaylov. 2017. “Understanding State Preferences with Text as Data: Introducing the UN General Debate Corpus.” Research &Amp; Politics 4 (April). https://doi.org/10.1177/2053168017712821.
Bauer, Matthias, Mateo Rojas-Carulla, Jakub Bartłomiej Świątkowski, Bernhard Schölkopf, and Richard E. Turner. 2017. “Discriminative k-Shot Learning Using Probabilistic Models.” arXiv. https://doi.org/10.48550/ARXIV.1706.00326.
Bayat, A. 2002. “Science, Medicine, and the Future: Bioinformatics.” BMJ 324 (April). https://doi.org/10.1136/bmj.324.7344.1018.
Baydin, Atilim Gunes, Barak A. Pearlmutter, Alexey Andreyevich Radul, and Jeffrey Mark Siskind. 2015. “Automatic Differentiation in Machine Learning: A Survey,” February. http://arxiv.org/abs/1502.05767v4.
Bayer, Justin, Christian Osendorfer, Daniela Korhammer, Nutan Chen, Sebastian Urban, and Patrick van der Smagt. 2013. “On Fast Dropout and Its Applicability to Recurrent Networks.” arXiv. https://doi.org/10.48550/ARXIV.1311.0701.
“Bayesian Statistics 7.” 2003, July. https://doi.org/10.1093/oso/9780198526155.001.0001.
Beam, Elizabeth, Christopher Potts, Russell A. Poldrack, and Amit Etkin. 2021. “A Data-Driven Framework for Mapping Domains of Human Neurobiology.” Nature Neuroscience 24 (November). https://doi.org/10.1038/s41593-021-00948-9.
Bean, James C. 1994. “Genetic Algorithms and Random Keys for Sequencing and Optimization.” ORSA Journal on Computing 6 (May). https://doi.org/10.1287/ijoc.6.2.154.
Beaumont, M. A., J.-M. Cornuet, J.-M. Marin, and C. P. Robert. 2009. “Adaptive Approximate Bayesian Computation.” Biometrika 96 (October). https://doi.org/10.1093/biomet/asp052.
Beck, Daniel, Gholamreza Haffari, and Trevor Cohn. 2018. “Graph-to-Sequence Learning Using Gated Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1806.09835.
Beck, Matthias, Sinai Robins, and Steven V Sam. 2009. “Positivity Theorems for Solid-Angle Polynomials.” arXiv. https://doi.org/10.48550/ARXIV.0906.4031.
Beck, Nathaniel, Gary King, and Langche Zeng. 2000. “Improving Quantitative Studies of International Conflict: A Conjecture.” American Political Science Review 94 (March). https://doi.org/10.2307/2586378.
Becker, Brett A., Paul Denny, James Finnie-Ansley, Andrew Luxton-Reilly, James Prather, and Eddie Antonio Santos. 2023. “Programming Is Hard - or at Least It Used to Be.” Proceedings of the 54th ACM Technical Symposium on Computer Science Education V. 1, March. https://doi.org/10.1145/3545945.3569759.
Bedford, Tim, and Roger M. Cooke. 2002. “Vines–a New Graphical Model for Dependent Random Variables.” The Annals of Statistics 30 (August). https://doi.org/10.1214/aos/1031689016.
Beffara, Vincent, and Hugo Duminil-Copin. 2011. “The Self-Dual Point of the Two-Dimensional Random-Cluster Model Is Critical for q ≥ 1.” Probability Theory and Related Fields 153 (March). https://doi.org/10.1007/s00440-011-0353-8.
Behera, Bibek, Manoj Joshi, Abhilash KK, and Mohammad Ansari Ismail. 2017. “Distributed Vector Representation of Shopping Items, the Customer and Shopping Cart to Build a Three Fold Recommendation System.” arXiv. https://doi.org/10.48550/ARXIV.1705.06338.
Behrmann, Jens, Will Grathwohl, Ricky T. Q. Chen, David Duvenaud, and Jörn-Henrik Jacobsen. 2018. “Invertible Residual Networks.” arXiv. https://doi.org/10.48550/ARXIV.1811.00995.
Bekkerman, Ron, and Matan Gavish. 2011. “High-Precision Phrase-Based Document Classification on a Modern Scale.” Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2020408.2020449.
Bekoulis, Giannis, Johannes Deleu, Thomas Demeester, and Chris Develder. 2018. “Adversarial Training for Multi-Context Joint Entity and Relation Extraction.” arXiv. https://doi.org/10.48550/ARXIV.1808.06876.
Belagiannis, Vasileios, Azade Farshad, and Fabio Galasso. 2018. “Adversarial Network Compression.” arXiv. https://doi.org/10.48550/ARXIV.1803.10750.
Beliaev, Stanislav, and Boris Ginsburg. 2021. “TalkNet 2: Non-Autoregressive Depth-Wise Separable Convolutional Model for Speech Synthesis with Explicit Pitch and Duration Prediction.” arXiv. https://doi.org/10.48550/ARXIV.2104.08189.
Belinkov, Yonatan, and James Glass. 2018. “Analysis Methods in Neural Language Processing: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.1812.08951.
Bellamy, Rachel K. E., Kuntal Dey, Michael Hind, Samuel C. Hoffman, Stephanie Houde, Kalapriya Kannan, Pranay Lohia, et al. 2018. “AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias.” arXiv. https://doi.org/10.48550/ARXIV.1810.01943.
Bellemare, Marc G., Will Dabney, and Rémi Munos. 2017. “A Distributional Perspective on Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1707.06887.
Bellet, Aurélien, Amaury Habrard, and Marc Sebban. 2013. “A Survey on Metric Learning for Feature Vectors and Structured Data.” arXiv. https://doi.org/10.48550/ARXIV.1306.6709.
Bello, Irwan, Hieu Pham, Quoc V. Le, Mohammad Norouzi, and Samy Bengio. 2016. “Neural Combinatorial Optimization with Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.09940.
Bello, Irwan, Barret Zoph, Ashish Vaswani, Jonathon Shlens, and Quoc V. Le. 2019. “Attention Augmented Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1904.09925.
Belrose, Nora, Zach Furman, Logan Smith, Danny Halawi, Igor Ostrovsky, Lev McKinney, Stella Biderman, and Jacob Steinhardt. 2023. “Eliciting Latent Predictions from Transformers with the Tuned Lens.” arXiv. https://doi.org/10.48550/ARXIV.2303.08112.
Beltagy, Iz, Matthew E. Peters, and Arman Cohan. 2020. “Longformer: The Long-Document Transformer,” April. http://arxiv.org/abs/2004.05150v2.
Bemis, Guy W., and Mark A. Murcko. 1996. “The Properties of Known Drugs. 1. Molecular Frameworks.” Journal of Medicinal Chemistry 39 (January). https://doi.org/10.1021/jm9602928.
Ben-David, Noa, and Sivan Sabato. 2021. “Active Structure Learning of Bayesian Networks in an Observational Setting.” arXiv. https://doi.org/10.48550/ARXIV.2103.13796.
Bender, Emily M., Timnit Gebru, Angelina McMillan-Major, and Shmargaret Shmitchell. 2021. “On the Dangers of Stochastic Parrots.” Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, March. https://doi.org/10.1145/3442188.3445922.
Bengio, Emmanuel, Pierre-Luc Bacon, Joelle Pineau, and Doina Precup. 2015. “Conditional Computation in Neural Networks for Faster Models.” arXiv. https://doi.org/10.48550/ARXIV.1511.06297.
Bengio, Samy, Oriol Vinyals, Navdeep Jaitly, and Noam Shazeer. 2015. “Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1506.03099.
Bengio, Y. 2009. “Learning Deep Architectures for AI.” Foundations and Trends® in Machine Learning 2. https://doi.org/10.1561/2200000006.
Bengio, Yoshua. 2012. “Practical Recommendations for Gradient-Based Training of Deep Architectures.” arXiv. https://doi.org/10.48550/ARXIV.1206.5533.
———. 2014. “How Auto-Encoders Could Provide Credit Assignment in Deep Networks via Target Propagation.” arXiv. https://doi.org/10.48550/ARXIV.1407.7906.
———. 2017. “The Consciousness Prior.” arXiv. https://doi.org/10.48550/ARXIV.1709.08568.
Bengio, Yoshua, Nicolas Boulanger-Lewandowski, and Razvan Pascanu. 2012. “Advances in Optimizing Recurrent Networks.” arXiv. https://doi.org/10.48550/ARXIV.1212.0901.
Bengio, Yoshua, Nicholas Léonard, and Aaron Courville. 2013. “Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation,” August. http://arxiv.org/abs/1308.3432v1.
Bengio, Yoshua, Andrea Lodi, and Antoine Prouvost. 2018. “Machine Learning for Combinatorial Optimization: A Methodological Tour d’horizon.” arXiv. https://doi.org/10.48550/ARXIV.1811.06128.
Bengio, Yoshua, Éric Thibodeau-Laufer, Guillaume Alain, and Jason Yosinski. 2013. “Deep Generative Stochastic Networks Trainable by Backprop,” June. http://arxiv.org/abs/1306.1091v5.
Bengio, Y., P. Simard, and P. Frasconi. 1994. “Learning Long-Term Dependencies with Gradient Descent Is Difficult.” IEEE Transactions on Neural Networks 5 (March). https://doi.org/10.1109/72.279181.
Benhenda, Mostapha. 2017. “ChemGAN Challenge for Drug Discovery: Can AI Reproduce Natural Chemical Diversity?” arXiv. https://doi.org/10.48550/ARXIV.1708.08227.
Benitez, Narciso. 2000. “Bayesian Photometric Redshift Estimation.” The Astrophysical Journal 536 (June). https://doi.org/10.1086/308947.
Benmeziane, Hadjer, Kaoutar El Maghraoui, Hamza Ouarnoughi, Smail Niar, Martin Wistuba, and Naigang Wang. 2021. “A Comprehensive Survey on Hardware-Aware Neural Architecture Search.” arXiv. https://doi.org/10.48550/ARXIV.2101.09336.
Ben-Michael, Eli, D. James Greiner, Kosuke Imai, and Zhichao Jiang. 2021. “Safe Policy Learning Through Extrapolation: Application to Pre-Trial Risk Assessment.” arXiv. https://doi.org/10.48550/ARXIV.2109.11679.
Benson, Alan, and Nial Friel. 2021. “Bayesian Inference, Model Selection and Likelihood Estimation Using Fast Rejection Sampling: The Conway-Maxwell-Poisson Distribution.” Bayesian Analysis 16 (September). https://doi.org/10.1214/20-ba1230.
Beraha, Mario, Alessandra Guglielmi, and Fernando A. Quintana. 2021. “The Semi-Hierarchical Dirichlet Process and Its Application to Clustering Homogeneous Distributions.” Bayesian Analysis 16 (December). https://doi.org/10.1214/21-ba1278.
Berenguer, Abel Díaz, Tanmoy Mukherjee, Matias Bossa, Nikos Deligiannis, and Hichem Sahli. 2022. “Representation Learning with Information Theory for COVID-19 Detection.” arXiv. https://doi.org/10.48550/ARXIV.2207.01437.
Berg, Rianne van den, Thomas N. Kipf, and Max Welling. 2017. “Graph Convolutional Matrix Completion.” arXiv. https://doi.org/10.48550/ARXIV.1706.02263.
Berger, Adam, and John Lafferty. 1999. “Information Retrieval as Statistical Translation.” Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, August. https://doi.org/10.1145/312624.312681.
Berger, Noam, and Marek Biskup. 2006. “Quenched Invariance Principle for Simple Random Walk on Percolation Clusters.” Probability Theory and Related Fields 137 (April). https://doi.org/10.1007/s00440-006-0498-z.
Bergstra, James, Olivier Breuleux, Frédéric Bastien, Pascal Lamblin, Razvan Pascanu, Guillaume Desjardins, Joseph Turian, David Warde-Farley, and Yoshua Bengio. 2010. “Theano: A CPU and GPU Math Compiler in Python.” Proceedings of the Python in Science Conference. https://doi.org/10.25080/majora-92bf1922-003.
Bermejo-Vega, Juan, and Maarten Van den Nest. 2012. “Classical Simulations of Abelian-Group Normalizer Circuits with Intermediate Measurements.” arXiv. https://doi.org/10.48550/ARXIV.1210.3637.
Bernstein, Daniel S., Robert Givan, Neil Immerman, and Shlomo Zilberstein. 2002. “The Complexity of Decentralized Control of Markov Decision Processes.” Mathematics of Operations Research 27 (November). https://doi.org/10.1287/moor.27.4.819.297.
Berrendorf, Max, Ludwig Wacker, and Evgeniy Faerman. 2021. “A Critical Assessment of State-of-the-Art in Entity Alignment.” Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-72240-1_2.
Berrone, Stefano, Claudio Canuto, and Moreno Pintore. 2022. “Variational Physics Informed Neural Networks: The Role of Quadratures and Test Functions.” Journal of Scientific Computing 92 (August). https://doi.org/10.1007/s10915-022-01950-4.
Bertasius, Gedas, Heng Wang, and Lorenzo Torresani. 2021. “Is Space-Time Attention All You Need for Video Understanding?” arXiv. https://doi.org/10.48550/ARXIV.2102.05095.
Berthelot, David, Nicholas Carlini, Ekin D. Cubuk, Alex Kurakin, Kihyuk Sohn, Han Zhang, and Colin Raffel. 2019. “ReMixMatch: Semi-Supervised Learning with Distribution Alignment and Augmentation Anchoring.” arXiv. https://doi.org/10.48550/ARXIV.1911.09785.
Berthelot, David, Thomas Schumm, and Luke Metz. 2017. “BEGAN: Boundary Equilibrium Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1703.10717.
Berti, Alessandro, Daniel Schuster, and Wil M. P. van der Aalst. 2023. “Abstractions, Scenarios, and Prompt Definitions for Process Mining with LLMs: A Case Study.” arXiv. https://doi.org/10.48550/ARXIV.2307.02194.
Bertin-Mahieux, Thierry, Daniel P. W. Ellis, Brian Whitman, and Paul Lamere. 2011. “The Million Song Dataset.” Columbia University. https://doi.org/10.7916/D8NZ8J07.
Bertsekas, Dimitri P. 2011. “Approximate Policy Iteration: A Survey and Some New Methods.” Journal of Control Theory and Applications 9 (July). https://doi.org/10.1007/s11768-011-1005-3.
———. 2015. “Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.1507.01030.
Bertsimas, Dimitris, Vishal Gupta, and Nathan Kallus. 2014a. “Data-Driven Robust Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1401.0212.
———. 2014b. “Robust Sample Average Approximation.” arXiv. https://doi.org/10.48550/ARXIV.1408.4445.
Bertucci, Donald, Md Montaser Hamid, Yashwanthi Anand, Anita Ruangrotsakun, Delyar Tabatabai, Melissa Perez, and Minsuk Kahng. 2022. “DendroMap: Visual Exploration of Large-Scale Image Datasets for Machine Learning with Treemaps.” arXiv. https://doi.org/10.48550/ARXIV.2205.06935.
Besag, Julian, and Peter J. Green. 1993. “Spatial Statistics and Bayesian Computation.” Journal of the Royal Statistical Society: Series B (Methodological) 55 (September). https://doi.org/10.1111/j.2517-6161.1993.tb01467.x.
Beskos, Alexandros, Natesh S. Pillai, Gareth O. Roberts, Jesus M. Sanz-Serna, and Andrew M. Stuart. 2010. “Optimal Tuning of the Hybrid Monte-Carlo Algorithm.” arXiv. https://doi.org/10.48550/ARXIV.1001.4460.
Besl, P. J., and Neil D. McKay. 1992. “A Method for Registration of 3-d Shapes.” IEEE Transactions on Pattern Analysis and Machine Intelligence 14 (February). https://doi.org/10.1109/34.121791.
Besse, Philippe, Brendan Guillouet, Jean-Michel Loubes, and Royer François. 2015. “Review and Perspective for Distance Based Trajectory Clustering.” arXiv. https://doi.org/10.48550/ARXIV.1508.04904.
Betancourt, Michael. 2017. “A Conceptual Introduction to Hamiltonian Monte Carlo.” arXiv. https://doi.org/10.48550/ARXIV.1701.02434.
Bethge, Joseph, Christian Bartz, Haojin Yang, Ying Chen, and Christoph Meinel. 2020. “MeliusNet: Can Binary Neural Networks Achieve MobileNet-Level Accuracy?” arXiv. https://doi.org/10.48550/ARXIV.2001.05936.
Bethge, Joseph, Haojin Yang, Marvin Bornstein, and Christoph Meinel. 2019. “Back to Simplicity: How to Train Accurate BNNs from Scratch?” arXiv. https://doi.org/10.48550/ARXIV.1906.08637.
Beyer, Lucas, Olivier J. Hénaff, Alexander Kolesnikov, Xiaohua Zhai, and Aäron van den Oord. 2020. “Are We Done with ImageNet?” arXiv. https://doi.org/10.48550/ARXIV.2006.07159.
Beygelzimer, Alina, Daniel Hsu, John Langford, and Tong Zhang. 2010. “Agnostic Active Learning Without Constraints.” arXiv. https://doi.org/10.48550/ARXIV.1006.2588.
Beygelzimer, Alina, Satyen Kale, and Haipeng Luo. 2015. “Optimal and Adaptive Algorithms for Online Boosting.” arXiv. https://doi.org/10.48550/ARXIV.1502.02651.
Bezanson, Jeff, Alan Edelman, Stefan Karpinski, and Viral B. Shah. 2014. “Julia: A Fresh Approach to Numerical Computing.” arXiv. https://doi.org/10.48550/ARXIV.1411.1607.
Bhagoji, Arjun Nitin, Daniel Cullina, Chawin Sitawarin, and Prateek Mittal. 2017. “Enhancing Robustness of Machine Learning Systems via Data Transformations.” arXiv. https://doi.org/10.48550/ARXIV.1704.02654.
Bhagoji, Arjun Nitin, Warren He, Bo Li, and Dawn Song. 2017. “Exploring the Space of Black-Box Attacks on Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1712.09491.
Bhalla, Sejal, Mayank Goel, and Rushil Khurana. 2021. “IMU2Doppler.” Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5 (December). https://doi.org/10.1145/3494994.
Bhandare, Aishwarya, Vamsi Sripathi, Deepthi Karkada, Vivek Menon, Sun Choi, Kushal Datta, and Vikram Saletore. 2019. “Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model.” arXiv. https://doi.org/10.48550/ARXIV.1906.00532.
Bharadiya, Jasmin Praful. 2023. “A Comprehensive Survey of Deep Learning Techniques Natural Language Processing.” European Journal of Technology 7 (May). https://doi.org/10.47672/ejt.1473.
Bhardwaj, Kartikeya, Naveen Suda, and Radu Marculescu. 2019. “Dream Distillation: A Data-Independent Model Compression Framework.” arXiv. https://doi.org/10.48550/ARXIV.1905.07072.
Bhatia, Sahil, and Rishabh Singh. 2016. “Automated Correction for Syntax Errors in Programming Assignments Using Recurrent Neural Networks,” March. http://arxiv.org/abs/1603.06129v1.
Bhatia, Shraey, Jey Han Lau, and Timothy Baldwin. 2016. “Automatic Labelling of Topics with Neural Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1612.05340.
Bhatia, Siddharth, Rui Liu, Bryan Hooi, Minji Yoon, Kijung Shin, and Christos Faloutsos. 2022. “Real-Time Anomaly Detection in Edge Streams.” ACM Transactions on Knowledge Discovery from Data 16 (January). https://doi.org/10.1145/3494564.
Bhatnagar, Aadyot, Paul Kassianik, Chenghao Liu, Tian Lan, Wenzhuo Yang, Rowan Cassius, Doyen Sahoo, et al. 2021. “Merlion: A Machine Learning Library for Time Series.” arXiv. https://doi.org/10.48550/ARXIV.2109.09265.
Bhattacharya, Kaushik, Bamdad Hosseini, Nikola B. Kovachki, and Andrew M. Stuart. 2021. “Model Reduction and Neural Networks for Parametric PDEs.” The SMAI Journal of Computational Mathematics 7 (July). https://doi.org/10.5802/smai-jcm.74.
Bhatti, Shehroze, Alban Desmaison, Ondrej Miksik, Nantas Nardelli, N. Siddharth, and Philip H. S. Torr. 2016. “Playing Doom with SLAM-Augmented Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1612.00380.
Bian, Ning, Xianpei Han, Le Sun, Hongyu Lin, Yaojie Lu, and Ben He. 2023. “ChatGPT Is a Knowledgeable but Inexperienced Solver: An Investigation of Commonsense Problem in Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.16421.
Bickel, Peter J., Nathan Boley, James B. Brown, Haiyan Huang, and Nancy R. Zhang. 2010. “Subsampling Methods for Genomic Inference.” The Annals of Applied Statistics 4 (December). https://doi.org/10.1214/10-aoas363.
Bickerton, G. Richard, Gaia V. Paolini, Jérémy Besnard, Sorel Muresan, and Andrew L. Hopkins. 2012. “Quantifying the Chemical Beauty of Drugs.” Nature Chemistry 4 (January). https://doi.org/10.1038/nchem.1243.
Biderman, Stella, USVSN Sai Prashanth, Lintang Sutawika, Hailey Schoelkopf, Quentin Anthony, Shivanshu Purohit, and Edward Raff. 2023. “Emergent and Predictable Memorization in Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.11158.
Bietti, Alberto, Vivien Cabannes, Diane Bouchacourt, Herve Jegou, and Leon Bottou. 2023. “Birth of a Transformer: A Memory Viewpoint.” arXiv. https://doi.org/10.48550/ARXIV.2306.00802.
Bietti, Alberto, Grégoire Mialon, Dexiong Chen, and Julien Mairal. 2018. “A Kernel Perspective for Regularizing Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1810.00363.
Biever, Celeste. 2023. “ChatGPT Broke the Turing Test — the Race Is on for New Ways to Assess AI.” Nature 619 (July). https://doi.org/10.1038/d41586-023-02361-7.
Bigdeli, Siavash Arjomand, and Matthias Zwicker. 2017. “Image Restoration Using Autoencoding Priors.” arXiv. https://doi.org/10.48550/ARXIV.1703.09964.
Biggio, Battista, Blaine Nelson, and Pavel Laskov. 2012. “Poisoning Attacks Against Support Vector Machines.” arXiv. https://doi.org/10.48550/ARXIV.1206.6389.
Biggio, Battista, and Fabio Roli. 2018. “Wild Patterns: Ten Years After the Rise of Adversarial Machine Learning.” Pattern Recognition 84 (December). https://doi.org/10.1016/j.patcog.2018.07.023.
Bin, Huang, Chen Weihai, Wu Xingming, and Lin Chun-Liang. 2017. “High-Quality Face Image SR Using Conditional Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1707.00737.
Birk, Alexander V., Wesley M. Chao, Shaoyi Liu, Yi Soong, and Hazel H. Szeto. 2015. “Disruption of Cytochrome c Heme Coordination Is Responsible for Mitochondrial Injury During Ischemia.” Biochimica Et Biophysica Acta (BBA) - Bioenergetics 1847 (October). https://doi.org/10.1016/j.bbabio.2015.06.006.
Birnie, Claire, Haithem Jarraya, and Fredrik Hansteen. 2021. “An Introduction to Distributed Training of Deep Neural Networks for Segmentation Tasks with Large Seismic Datasets.” arXiv. https://doi.org/10.48550/ARXIV.2102.13003.
Bishop, Lea. 2023. “A Computer Wrote This Paper: What ChatGPT Means for Education, Research, and Writing.” SSRN Electronic Journal. https://doi.org/10.2139/ssrn.4338981.
Biswas, Russa, Lucie-Aimée Kaffee, Michael Cochez, Stefania Dumbrava, Theis E. Jendal, Matteo Lissandrini, Vanessa Lopez, et al. 2023. “Knowledge Graph Embeddings: Open Challenges and Opportunities.” Schloss Dagstuhl – Leibniz-Zentrum Für Informatik. https://doi.org/10.4230/TGDK.1.1.4.
Bitzenbauer, Philipp. 2023. “ChatGPT in Physics Education: A Pilot Study on Easy-to-Implement Activities.” Contemporary Educational Technology 15 (July). https://doi.org/10.30935/cedtech/13176.
Blaauw, Merlijn, and Jordi Bonada. 2017. “A Neural Parametric Singing Synthesizer.” arXiv. https://doi.org/10.48550/ARXIV.1704.03809.
Blair-Stanek, Andrew, Nils Holzenberger, and Benjamin Van Durme. 2023. “Can GPT-3 Perform Statutory Reasoning?” arXiv. https://doi.org/10.48550/ARXIV.2302.06100.
Blanchard, G., C. Schäfer, Y. Rozenholc, and K.-R. Müller. 2007. “Optimal Dyadic Decision Trees.” Machine Learning 66 (January). https://doi.org/10.1007/s10994-007-0717-6.
Blanchet, Juliette, and Anthony C. Davison. 2011. “Spatial Modeling of Extreme Snow Depth.” The Annals of Applied Statistics 5 (September). https://doi.org/10.1214/11-aoas464.
Blei, David M., Alp Kucukelbir, and Jon D. McAuliffe. 2017. “Variational Inference: A Review for Statisticians.” Journal of the American Statistical Association 112 (April). https://doi.org/10.1080/01621459.2017.1285773.
Blei, David M., and John D. Lafferty. 2007. “A Correlated Topic Model of Science.” The Annals of Applied Statistics 1 (June). https://doi.org/10.1214/07-aoas114.
Blockeel, Hendrik, Luc De Raedt, and Jan Ramon. 2000. “Top-down Induction of Clustering Trees.” arXiv. https://doi.org/10.48550/ARXIV.CS/0011032.
Blondel, Vincent D, Jean-Loup Guillaume, Renaud Lambiotte, and Etienne Lefebvre. 2008. “Fast Unfolding of Communities in Large Networks.” Journal of Statistical Mechanics: Theory and Experiment 2008 (October). https://doi.org/10.1088/1742-5468/2008/10/p10008.
Bloom, Burton H. 1970. “Space/Time Trade-Offs in Hash Coding with Allowable Errors.” Communications of the ACM 13 (July). https://doi.org/10.1145/362686.362692.
Blot, Michael, David Picard, Matthieu Cord, and Nicolas Thome. 2016. “Gossip Training for Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.09726.
Bluche, Théodore, and Thibault Gisselbrecht. 2020. “Predicting Detection Filters for Small Footprint Open-Vocabulary Keyword Spotting.” Interspeech 2020, October. https://doi.org/10.21437/interspeech.2020-1186.
Blum, Avrim, and Tom Mitchell. 1998. “Combining Labeled and Unlabeled Data with Co-Training.” Proceedings of the Eleventh Annual Conference on Computational Learning Theory, July. https://doi.org/10.1145/279943.279962.
Blumer, Anselm, A. Ehrenfeucht, David Haussler, and Manfred K. Warmuth. 1989. “Learnability and the Vapnik-Chervonenkis Dimension.” Journal of the ACM 36 (October). https://doi.org/10.1145/76359.76371.
Blundell, Charles, Julien Cornebise, Koray Kavukcuoglu, and Daan Wierstra. 2015. “Weight Uncertainty in Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1505.05424.
Blundell, Charles, Yee Whye Teh, and Katherine A. Heller. 2012. “Bayesian Rose Trees.” arXiv. https://doi.org/10.48550/ARXIV.1203.3468.
Bodnar, Olha, and Taras Bodnar. 2023. “Objective Bayesian Meta-Analysis Based on Generalized Marginal Multivariate Random Effects Model.” Bayesian Analysis -1 (January). https://doi.org/10.1214/23-ba1363.
Boella, Guido, and Leendert van der Torre. 2007. “Institutions with a Hierarchy of Authorities in Distributed Dynamic Environments.” Artificial Intelligence and Law 16 (October). https://doi.org/10.1007/s10506-007-9059-8.
Bogdan, Krzysztof, Krzysztof Burdzy, and Zhen-Qing Chen. 2003. “Censored Stable Processes.” Probability Theory and Related Fields 127 (May). https://doi.org/10.1007/s00440-003-0275-1.
Bogin, Ben, Matt Gardner, and Jonathan Berant. 2019. “Representing Schema Structure with Graph Neural Networks for Text-to-SQL Parsing.” arXiv. https://doi.org/10.48550/ARXIV.1905.06241.
Bogolin, Simion-Vlad, Ioana Croitoru, Hailin Jin, Yang Liu, and Samuel Albanie. 2021. “Cross Modal Retrieval with Querybank Normalisation.” arXiv. https://doi.org/10.48550/ARXIV.2112.12777.
Böhle, Moritz, Mario Fritz, and Bernt Schiele. 2022. “B-Cos Networks: Alignment Is All We Need for Interpretability.” arXiv. https://doi.org/10.48550/ARXIV.2205.10268.
Bohlin, R., and L. E. Kavraki. 2000. “Path Planning Using Lazy PRM.” Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065). https://doi.org/10.1109/robot.2000.844107.
Boiko, Daniil A., Robert MacKnight, and Gabe Gomes. 2023. “Emergent Autonomous Scientific Research Capabilities of Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.05332.
Boillet, Mélodie, Marie-Laurence Bonhomme, Dominique Stutzmann, and Christopher Kermorvant. 2019. “HORAE.” Proceedings of the 5th International Workshop on Historical Document Imaging and Processing, September. https://doi.org/10.1145/3352631.3352633.
Bojanowski, Piotr, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2016. “Enriching Word Vectors with Subword Information.” arXiv. https://doi.org/10.48550/ARXIV.1607.04606.
Bojchevski, Aleksandar, and Stephan Günnemann. 2017. “Deep Gaussian Embedding of Graphs: Unsupervised Inductive Learning via Ranking.” arXiv. https://doi.org/10.48550/ARXIV.1707.03815.
Bolin, David, and Finn Lindgren. 2011. “Spatial Models Generated by Nested Stochastic Partial Differential Equations, with an Application to Global Ozone Mapping.” The Annals of Applied Statistics 5 (March). https://doi.org/10.1214/10-aoas383.
Bolukbasi, Tolga, Kai-Wei Chang, James Zou, Venkatesh Saligrama, and Adam Kalai. 2016. “Man Is to Computer Programmer as Woman Is to Homemaker? Debiasing Word Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1607.06520.
Bommineni, Vikas L, Sanaea Bhagwagar, Daniel Balcarcel, Christos Davatzikos, and Donald Boyer. 2023. “Performance of ChatGPT on the MCAT: The Road to Personalized and Equitable Premedical Learning,” March. https://doi.org/10.1101/2023.03.05.23286533.
Boniol, Paul, and Themis Palpanas. 2020. “Series2Graph.” Proceedings of the VLDB Endowment 13 (August). https://doi.org/10.14778/3407790.3407792.
Boniol, Paul, John Paparrizos, Themis Palpanas, and Michael J. Franklin. 2021. “SAND.” Proceedings of the VLDB Endowment 14 (June). https://doi.org/10.14778/3467861.3467863.
Boom, Willem van den, Maria De Iorio, and Alexandros Beskos. 2023. “Bayesian Learning of Graph Substructures.” Bayesian Analysis 18 (December). https://doi.org/10.1214/22-ba1338.
Bordes, Antoine, Y-Lan Boureau, and Jason Weston. 2016. “Learning End-to-End Goal-Oriented Dialog.” arXiv. https://doi.org/10.48550/ARXIV.1605.07683.
Bordes, Antoine, Sumit Chopra, and Jason Weston. 2014. “Question Answering with Subgraph Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1406.3676.
Bordes, Antoine, Nicolas Usunier, Sumit Chopra, and Jason Weston. 2015. “Large-Scale Simple Question Answering with Memory Networks.” arXiv. https://doi.org/10.48550/ARXIV.1506.02075.
Bordes, Antoine, Jason Weston, Ronan Collobert, and Yoshua Bengio. 2011. “Learning Structured Embeddings of Knowledge Bases.” Proceedings of the AAAI Conference on Artificial Intelligence 25 (August). https://doi.org/10.1609/aaai.v25i1.7917.
Bordes, Antoine, Jason Weston, and Nicolas Usunier. 2014. “Open Question Answering with Weakly Supervised Embedding Models.” arXiv. https://doi.org/10.48550/ARXIV.1404.4326.
Bordin, Nicola, Ian Sillitoe, Vamsi Nallapareddy, Clemens Rauer, Su Datt Lam, Vaishali P. Waman, Neeladri Sen, et al. 2023. “AlphaFold2 Reveals Commonalities and Novelties in Protein Structure Space for 21 Model Organisms.” Communications Biology 6 (February). https://doi.org/10.1038/s42003-023-04488-9.
Borges, Francisco, Georgios Balikas, Marc Brette, Guillaume Kempf, Arvind Srikantan, Matthieu Landos, Darya Brazouskaya, and Qianqian Shi. 2020. “Query Understanding for Natural Language Enterprise Search.” arXiv. https://doi.org/10.48550/ARXIV.2012.06238.
Borgwardt, K. M., C. S. Ong, S. Schonauer, S. V. N. Vishwanathan, A. J. Smola, and H.-P. Kriegel. 2005. “Protein Function Prediction via Graph Kernels.” Bioinformatics 21 (June). https://doi.org/10.1093/bioinformatics/bti1007.
Borodin, Alexei, and Ivan Corwin. 2013. “Macdonald Processes.” Probability Theory and Related Fields 158 (March). https://doi.org/10.1007/s00440-013-0482-3.
Borsos, Zalán, Raphaël Marinier, Damien Vincent, Eugene Kharitonov, Olivier Pietquin, Matt Sharifi, Dominik Roblek, et al. 2022. “AudioLM: A Language Modeling Approach to Audio Generation.” arXiv. https://doi.org/10.48550/ARXIV.2209.03143.
Bosselut, Antoine, Hannah Rashkin, Maarten Sap, Chaitanya Malaviya, Asli Celikyilmaz, and Yejin Choi. 2019. “COMET: Commonsense Transformers for Automatic Knowledge Graph Construction.” arXiv. https://doi.org/10.48550/ARXIV.1906.05317.
Botchkarev, Alexei. 2019. “A New Typology Design of Performance Metrics to Measure Errors in Machine Learning Regression Algorithms.” Interdisciplinary Journal of Information, Knowledge, and Management 14. https://doi.org/10.28945/4184.
Botha, Imke, Robert Kohn, and Christopher Drovandi. 2021. “Particle Methods for Stochastic Differential Equation Mixed Effects Models.” Bayesian Analysis 16 (June). https://doi.org/10.1214/20-ba1216.
Bottou, Léon, Frank E. Curtis, and Jorge Nocedal. 2016. “Optimization Methods for Large-Scale Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.04838.
Bouckaert, Remco, Joseph Heled, Denise Kühnert, Tim Vaughan, Chieh-Hsi Wu, Dong Xie, Marc A. Suchard, Andrew Rambaut, and Alexei J. Drummond. 2014. “BEAST 2: A Software Platform for Bayesian Evolutionary Analysis.” PLoS Computational Biology 10 (April). https://doi.org/10.1371/journal.pcbi.1003537.
Bouktif, Salah, Ali Fiaz, Ali Ouni, and Mohamed Serhani. 2018. “Optimal Deep Learning LSTM Model for Electric Load Forecasting Using Feature Selection and Genetic Algorithm: Comparison with Machine Learning Approaches †.” Energies 11 (June). https://doi.org/10.3390/en11071636.
Boulanger-Lewandowski, Nicolas, Yoshua Bengio, and Pascal Vincent. 2012. “Modeling Temporal Dependencies in High-Dimensional Sequences: Application to Polyphonic Music Generation and Transcription.” arXiv. https://doi.org/10.48550/ARXIV.1206.6392.
Boularias, Abdeslam, Oliver Kroemer, and Jan Peters. 2011. “Learning Robot Grasping from 3-d Images with Markov Random Fields.” 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, September. https://doi.org/10.1109/iros.2011.6094888.
Boulch, Alexandre. 2017. “ShaResNet: Reducing Residual Network Parameter Number by Sharing Weights.” arXiv. https://doi.org/10.48550/ARXIV.1702.08782.
Bourgade, Paul, László Erdös, and Horng-Tzer Yau. 2014. “Edge Universality of Beta Ensembles.” Communications in Mathematical Physics 332 (August). https://doi.org/10.1007/s00220-014-2120-z.
Bourgault, F., T. Furukawa, and H. F. Durrant-Whyte. n.d. “Coordinated Decentralized Search for a Lost Target in a Bayesian World.” Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453). https://doi.org/10.1109/iros.2003.1250604.
Bourtoule, Lucas, Varun Chandrasekaran, Christopher A. Choquette-Choo, Hengrui Jia, Adelin Travers, Baiwu Zhang, David Lie, and Nicolas Papernot. 2019. “Machine Unlearning.” arXiv. https://doi.org/10.48550/ARXIV.1912.03817.
Bousmalis, Konstantinos, Alex Irpan, Paul Wohlhart, Yunfei Bai, Matthew Kelcey, Mrinal Kalakrishnan, Laura Downs, et al. 2017. “Using Simulation and Domain Adaptation to Improve Efficiency of Deep Robotic Grasping.” arXiv. https://doi.org/10.48550/ARXIV.1709.07857.
Bousmalis, Konstantinos, Nathan Silberman, David Dohan, Dumitru Erhan, and Dilip Krishnan. 2016. “Unsupervised Pixel-Level Domain Adaptation with Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1612.05424.
Bouton, Maxime, Hasan Farooq, Julien Forgeat, Shruti Bothe, Meral Shirazipour, and Per Karlsson. 2021. “Coordinated Reinforcement Learning for Optimizing Mobile Networks.” arXiv. https://doi.org/10.48550/ARXIV.2109.15175.
Bouveyron, Charles, Etienne Côme, and Julien Jacques. 2015. “The Discriminative Functional Mixture Model for a Comparative Analysis of Bike Sharing Systems.” The Annals of Applied Statistics 9 (December). https://doi.org/10.1214/15-aoas861.
Bowers, Jeffrey S. 2000. “In Defense of Abstractionist Theories of Repetition Priming and Word Identification.” Psychonomic Bulletin &Amp; Review 7 (March). https://doi.org/10.3758/bf03210726.
Bowman, Samuel R. 2023. “Eight Things to Know about Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.00612.
Bowman, Samuel R., Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. “A Large Annotated Corpus for Learning Natural Language Inference.” arXiv. https://doi.org/10.48550/ARXIV.1508.05326.
Bowman, Samuel R., Luke Vilnis, Oriol Vinyals, Andrew M. Dai, Rafal Jozefowicz, and Samy Bengio. 2015. “Generating Sentences from a Continuous Space.” arXiv. https://doi.org/10.48550/ARXIV.1511.06349.
Boyer, Christopher N., and B. Wade Brorsen. 2013. “Implications of a Reserve Price in an Agent-Based Common-Value Auction.” Computational Economics 43 (December). https://doi.org/10.1007/s10614-013-9413-8.
Bozkurt, Aras, and Ramesh C. Sharma. 2023. “Generative AI and Prompt Engineering: The Art of Whispering to Let the Genie Out of the Algorithmic World,” July. https://doi.org/10.5281/ZENODO.8174941.
Bradbury, James, Stephen Merity, Caiming Xiong, and Richard Socher. 2016. “Quasi-Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.01576.
Bradshaw, John, Alexander G. de G. Matthews, and Zoubin Ghahramani. 2017. “Adversarial Examples, Uncertainty, and Transfer Testing Robustness in Gaussian Process Hybrid Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1707.02476.
Brainard, Jeffrey. 2023. “Journals Take up Arms Against AI-Written Text.” Science 379 (February). https://doi.org/10.1126/science.adh2762.
Bräm, Dominic Stefan, Neil Parrott, Lucy Hutchinson, and Bernhard Steiert. 2022. “Introduction of an Artificial Neural Network–Based Method for Concentration‐time Predictions.” CPT: Pharmacometrics &Amp; Systems Pharmacology 11 (May). https://doi.org/10.1002/psp4.12786.
Bran, Andres M, Sam Cox, Oliver Schilter, Carlo Baldassari, Andrew D White, and Philippe Schwaller. 2023. “ChemCrow: Augmenting Large-Language Models with Chemistry Tools.” arXiv. https://doi.org/10.48550/ARXIV.2304.05376.
Branco, Paula, Luis Torgo, and Rita Ribeiro. 2015. “A Survey of Predictive Modelling Under Imbalanced Distributions.” arXiv. https://doi.org/10.48550/ARXIV.1505.01658.
Branson, Steve, Grant Van Horn, Serge Belongie, and Pietro Perona. 2014. “Bird Species Categorization Using Pose Normalized Deep Convolutional Nets.” arXiv. https://doi.org/10.48550/ARXIV.1406.2952.
Brazdil, Pavel, and Christophe Giraud-Carrier. 2017. “Metalearning and Algorithm Selection: Progress, State of the Art and Introduction to the 2018 Special Issue.” Machine Learning 107 (December). https://doi.org/10.1007/s10994-017-5692-y.
Breiman, Leo, and Jerome H. Friedman. 1985. “Estimating Optimal Transformations for Multiple Regression and Correlation.” Journal of the American Statistical Association 80 (September). https://doi.org/10.1080/01621459.1985.10478157.
Bressem, Keno K., Jens-Michalis Papaioannou, Paul Grundmann, Florian Borchert, Lisa C. Adams, Leonhard Liu, Felix Busch, et al. 2024. “medBERT.de: A Comprehensive German BERT Model for the Medical Domain.” Expert Systems with Applications 237 (March). https://doi.org/10.1016/j.eswa.2023.121598.
Bresson, Xavier, and Thomas Laurent. 2017. “Residual Gated Graph ConvNets.” arXiv. https://doi.org/10.48550/ARXIV.1711.07553.
Bretó, Carles, Daihai He, Edward L. Ionides, and Aaron A. King. 2009. “Time Series Analysis via Mechanistic Models.” The Annals of Applied Statistics 3 (March). https://doi.org/10.1214/08-aoas201.
Breuer, Adam, Roee Eilat, and Udi Weinsberg. 2020. “Friend or Faux: Graph-Based Early Detection of Fake Accounts on Social Networks.” arXiv. https://doi.org/10.48550/ARXIV.2004.04834.
Briand, Philippe, and Ying Hu. 2006. “BSDE with Quadratic Growth and Unbounded Terminal Value.” Probability Theory and Related Fields 136 (April). https://doi.org/10.1007/s00440-006-0497-0.
———. 2007. “Quadratic BSDEs with Convex Generators and Unbounded Terminal Conditions.” Probability Theory and Related Fields 141 (August). https://doi.org/10.1007/s00440-007-0093-y.
Bridle, Sarah, Mandeep Gill, Alan Heavens, Catherine Heymans, F. William High, Henk Hoekstra, Mike Jarvis, et al. 2009. “Handbook for the GREAT08 Challenge: An Image Analysis Competition for Cosmological Lensing.” The Annals of Applied Statistics 3 (March). https://doi.org/10.1214/08-aoas222.
Briers, Mark, Arnaud Doucet, and Simon Maskell. 2009. “Smoothing Algorithms for State–Space Models.” Annals of the Institute of Statistical Mathematics 62 (June). https://doi.org/10.1007/s10463-009-0236-2.
Brinkmann, Levin, Fabian Baumann, Jean-François Bonnefon, Maxime Derex, Thomas F. Müller, Anne-Marie Nussberger, Agnieszka Czaplicka, et al. 2023. “Machine Culture.” Nature Human Behaviour 7 (November). https://doi.org/10.1038/s41562-023-01742-2.
Briola, Antonio, Jeremy Turiel, Riccardo Marcaccioli, Alvaro Cauderan, and Tomaso Aste. 2021. “Deep Reinforcement Learning for Active High Frequency Trading.” arXiv. https://doi.org/10.48550/ARXIV.2101.07107.
Britz, Denny, Anna Goldie, Minh-Thang Luong, and Quoc Le. 2017. “Massive Exploration of Neural Machine Translation Architectures.” arXiv. https://doi.org/10.48550/ARXIV.1703.03906.
Brochu, Eric, Vlad M. Cora, and Nando de Freitas. 2010. “A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1012.2599.
Brock, Andrew, Jeff Donahue, and Karen Simonyan. 2018. “Large Scale GAN Training for High Fidelity Natural Image Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1809.11096.
Brock, Andrew, Theodore Lim, J. M. Ritchie, and Nick Weston. 2016. “Neural Photo Editing with Introspective Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1609.07093.
———. 2017. “SMASH: One-Shot Model Architecture Search Through HyperNetworks.” arXiv. https://doi.org/10.48550/ARXIV.1708.05344.
Brockschmidt, Marc, Miltiadis Allamanis, Alexander L. Gaunt, and Oleksandr Polozov. 2018. “Generative Code Modeling with Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1805.08490.
Broder, A. Z. n.d. “On the Resemblance and Containment of Documents.” Proceedings. Compression and Complexity of SEQUENCES 1997 (Cat. No.97TB100171). https://doi.org/10.1109/sequen.1997.666900.
Brohan, Anthony, Noah Brown, Justice Carbajal, Yevgen Chebotar, Xi Chen, Krzysztof Choromanski, Tianli Ding, et al. 2023. “RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control.” arXiv. https://doi.org/10.48550/ARXIV.2307.15818.
Brown, Nathan, Marco Fiscato, Marwin H. S. Segler, and Alain C. Vaucher. 2019. “GuacaMol: Benchmarking Models for de Novo Molecular Design.” Journal of Chemical Information and Modeling 59 (March). https://doi.org/10.1021/acs.jcim.8b00839.
Brown, Noam, and Tuomas Sandholm. 2019. “Superhuman AI for Multiplayer Poker.” Science 365 (August). https://doi.org/10.1126/science.aay2400.
Brown, Tom B., Dandelion Mané, Aurko Roy, Martín Abadi, and Justin Gilmer. 2017. “Adversarial Patch.” arXiv. https://doi.org/10.48550/ARXIV.1712.09665.
Brown, Tom B., Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, et al. 2020. “Language Models Are Few-Shot Learners.” arXiv. https://doi.org/10.48550/ARXIV.2005.14165.
Bruna, Joan, Wojciech Zaremba, Arthur Szlam, and Yann LeCun. 2013. “Spectral Networks and Locally Connected Networks on Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1312.6203.
Brundage, Miles, Shahar Avin, Jack Clark, Helen Toner, Peter Eckersley, Ben Garfinkel, Allan Dafoe, et al. 2018. “The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation.” arXiv. https://doi.org/10.48550/ARXIV.1802.07228.
Bruni, Renato. 2004. “On Exact Selection of Minimally Unsatisfiable Subformulae.” Annals of Mathematics and Artificial Intelligence 43 (December). https://doi.org/10.1007/s10472-005-0418-4.
Bu, Jiajun, Shulong Tan, Chun Chen, Can Wang, Hao Wu, Lijun Zhang, and Xiaofei He. 2010. “Music Recommendation by Unified Hypergraph.” Proceedings of the 18th ACM International Conference on Multimedia, October. https://doi.org/10.1145/1873951.1874005.
Bubeck, Sébastien, Rémi Munos, Gilles Stoltz, and Csaba Szepesvari. 2010. “X-Armed Bandits.” arXiv. https://doi.org/10.48550/ARXIV.1001.4475.
Bubeck, Sébastien, Eric Price, and Ilya Razenshteyn. 2018. “Adversarial Examples from Computational Constraints.” arXiv. https://doi.org/10.48550/ARXIV.1805.10204.
Buchholz, Alexander, Nicolas Chopin, and Pierre E. Jacob. 2021. “Adaptive Tuning of Hamiltonian Monte Carlo Within Sequential Monte Carlo.” Bayesian Analysis 16 (September). https://doi.org/10.1214/20-ba1222.
Buciluǎ, Cristian, Rich Caruana, and Alexandru Niculescu-Mizil. 2006. “Model Compression.” Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/1150402.1150464.
Budak, Ceren, Sharad Goel, and Justin M. Rao. 2016. “Fair and Balanced? Quantifying Media Bias Through Crowdsourced Content Analysis.” Public Opinion Quarterly 80. https://doi.org/10.1093/poq/nfw007.
Budzianowski, Paweł, Stefan Ultes, Pei-Hao Su, Nikola Mrkšić, Tsung-Hsien Wen, Iñigo Casanueva, Lina Rojas-Barahona, and Milica Gašić. 2017. “Sub-Domain Modelling for Dialogue Management with Hierarchical Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1706.06210.
Budzianowski, Paweł, and Ivan Vulić. 2019. “Hello, It’s GPT-2 – How Can i Help You? Towards the Use of Pretrained Language Models for Task-Oriented Dialogue Systems.” arXiv. https://doi.org/10.48550/ARXIV.1907.05774.
Buehler, Markus J. 2023. “MechGPT, a Language-Based Strategy for Mechanics and Materials Modeling That Connects Knowledge Across Scales, Disciplines and Modalities.” arXiv. https://doi.org/10.48550/ARXIV.2310.10445.
Bui, Tien-Cuong, Van-Duc Le, Wen-Syan Li, and Sang Kyun Cha. 2022. “INGREX: An Interactive Explanation Framework for Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2211.01548.
Bulacu, Marius, Axel Brink, Tijn van der Zant, and Lambert Schomaker. 2009. “Recognition of Handwritten Numerical Fields in a Large Single-Writer Historical Collection.” 2009 10th International Conference on Document Analysis and Recognition. https://doi.org/10.1109/icdar.2009.8.
Bulatov, Aydar, Yuri Kuratov, Yermek Kapushev, and Mikhail S. Burtsev. 2023. “Scaling Transformer to 1M Tokens and Beyond with RMT.” arXiv. https://doi.org/10.48550/ARXIV.2304.11062.
Bulatov, Konstantin, Boris Savelyev, and Vladimir V. Arlazarov. 2020. “Next Integrated Result Modelling for Stopping the Text Field Recognition Process in a Video Using a Result Model with Per-Character Alternatives.” Twelfth International Conference on Machine Vision (ICMV 2019), January. https://doi.org/10.1117/12.2559447.
Bull, Adam D. 2011. “Convergence Rates of Efficient Global Optimization Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.1101.3501.
Buonaguidi, Bruno, Antonietta Mira, Herbert Bucheli, and Viton Vitanis. 2022. “Bayesian Quickest Detection of Credit Card Fraud.” Bayesian Analysis 17 (March). https://doi.org/10.1214/20-ba1254.
Burke, Robin, Bamshad Mobasher, Chad Williams, and Runa Bhaumik. 2006. “Classification Features for Attack Detection in Collaborative Recommender Systems.” Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/1150402.1150465.
Burkholder, Kai, Kenny Kwock, Yuesheng Xu, Jiaxin Liu, Chao Chen, and Sihong Xie. 2021. “Certification and Trade-Off of Multiple Fairness Criteria in Graph-Based Spam Detection.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3482325.
Bürkner, Paul-Christian. 2017. “Advanced Bayesian Multilevel Modeling with the r Package Brms.” arXiv. https://doi.org/10.48550/ARXIV.1705.11123.
Bürkner, Paul-Christian, Jonah Gabry, and Aki Vehtari. 2020. “Approximate Leave-Future-Out Cross-Validation for Bayesian Time Series Models.” Journal of Statistical Computation and Simulation 90 (June). https://doi.org/10.1080/00949655.2020.1783262.
Burns, Collin, Jesse Thomason, and Wesley Tansey. 2020. “Interpreting Black Box Models via Hypothesis Testing.” Proceedings of the 2020 ACM-IMS on Foundations of Data Science Conference, October. https://doi.org/10.1145/3412815.3416889.
Burns, Rod, John Lawson, Duncan McBain, and Daniel Soutar. 2019. “Accelerated Neural Networks on OpenCL Devices Using SYCL-DNN.” Proceedings of the International Workshop on OpenCL, May. https://doi.org/10.1145/3318170.3318183.
Busch, Kiran, Alexander Rochlitzer, Diana Sola, and Henrik Leopold. 2023. “Just Tell Me: Prompt Engineering in Business Process Management.” arXiv. https://doi.org/10.48550/ARXIV.2304.07183.
Butt, Asad A., and Robert T. Collins. 2013. “Multi-Target Tracking by Lagrangian Relaxation to Min-Cost Network Flow.” 2013 IEEE Conference on Computer Vision and Pattern Recognition, June. https://doi.org/10.1109/cvpr.2013.241.
Caccia, Massimo, Lucas Caccia, William Fedus, Hugo Larochelle, Joelle Pineau, and Laurent Charlin. 2018. “Language GANs Falling Short.” arXiv. https://doi.org/10.48550/ARXIV.1811.02549.
Caesar, Holger, Jasper Uijlings, and Vittorio Ferrari. 2016. “COCO-Stuff: Thing and Stuff Classes in Context.” arXiv. https://doi.org/10.48550/ARXIV.1612.03716.
Cai, Han, Chuang Gan, Tianzhe Wang, Zhekai Zhang, and Song Han. 2019. “Once-for-All: Train One Network and Specialize It for Efficient Deployment.” arXiv. https://doi.org/10.48550/ARXIV.1908.09791.
Cai, Han, Chuang Gan, Ligeng Zhu, and Song Han. 2020. “TinyTL: Reduce Activations, Not Trainable Parameters for Efficient on-Device Learning.” arXiv. https://doi.org/10.48550/ARXIV.2007.11622.
Cai, Han, Ligeng Zhu, and Song Han. 2018. “ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware.” arXiv. https://doi.org/10.48550/ARXIV.1812.00332.
Cai, Hongyun, Vincent W. Zheng, and Kevin Chen-Chuan Chang. 2017. “A Comprehensive Survey of Graph Embedding: Problems, Techniques and Applications.” arXiv. https://doi.org/10.48550/ARXIV.1709.07604.
Cai, Jian-Feng, Emmanuel J. Candes, and Zuowei Shen. 2008. “A Singular Value Thresholding Algorithm for Matrix Completion.” arXiv. https://doi.org/10.48550/ARXIV.0810.3286.
Cai, Jonathon, Richard Shin, and Dawn Song. 2017. “Making Neural Programming Architectures Generalize via Recursion.” arXiv. https://doi.org/10.48550/ARXIV.1704.06611.
Cai, Li. 2009. “High-Dimensional Exploratory Item Factor Analysis by a Metropolis–Hastings Robbins–Monro Algorithm.” Psychometrika 75 (July). https://doi.org/10.1007/s11336-009-9136-x.
Cai, Shengze, Zhicheng Wang, Sifan Wang, Paris Perdikaris, and George Em Karniadakis. 2021. “Physics-Informed Neural Networks for Heat Transfer Problems.” Journal of Heat Transfer 143 (April). https://doi.org/10.1115/1.4050542.
Cai, T. Tony, Tengyuan Liang, and Harrison H. Zhou. 2015. “Law of Log Determinant of Sample Covariance Matrix and Optimal Estimation of Differential Entropy for High-Dimensional Gaussian Distributions.” Journal of Multivariate Analysis 137 (May). https://doi.org/10.1016/j.jmva.2015.02.003.
Cai, Yucheng, Wentao Ma, Yuchuan Wu, Shuzheng Si, Yuan Shao, Zhijian Ou, and Yongbin Li. 2023. “UniPCM: Universal Pre-Trained Conversation Model with Task-Aware Automatic Prompt.” arXiv. https://doi.org/10.48550/ARXIV.2309.11065.
Cai, Zhaowei, Xiaodong He, Jian Sun, and Nuno Vasconcelos. 2017. “Deep Learning with Low Precision by Half-Wave Gaussian Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1702.00953.
Cai, Zhaowei, and Nuno Vasconcelos. 2019. “Cascade r-CNN: High Quality Object Detection and Instance Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1906.09756.
Calandra, Roberto, Jan Peters, Carl Edward Rasmussen, and Marc Peter Deisenroth. 2014. “Manifold Gaussian Processes for Regression.” arXiv. https://doi.org/10.48550/ARXIV.1402.5876.
Caldas, Sebastian, Sai Meher Karthik Duddu, Peter Wu, Tian Li, Jakub Konečný, H. Brendan McMahan, Virginia Smith, and Ameet Talwalkar. 2018. “LEAF: A Benchmark for Federated Settings.” arXiv. https://doi.org/10.48550/ARXIV.1812.01097.
Calderon-Ramirez, Saul, Armaghan Moemeni, David Elizondo, Simon Colreavy-Donnelly, Luis Fernando Chavarria-Estrada, and Miguel A. Molina-Cabello. 2020. “Correcting Data Imbalance for Semi-Supervised Covid-19 Detection Using x-Ray Chest Images.” arXiv. https://doi.org/10.48550/ARXIV.2008.08496.
Caliskan, Aylin, Joanna J. Bryson, and Arvind Narayanan. 2017. “Semantics Derived Automatically from Language Corpora Contain Human-Like Biases.” Science 356 (April). https://doi.org/10.1126/science.aal4230.
Calzavara, Stefano, Claudio Lucchese, and Gabriele Tolomei. 2019. “Adversarial Training of Gradient-Boosted Decision Trees.” Proceedings of the 28th ACM International Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3357384.3358149.
Camacho-Collados, Jose, and Mohammad Taher Pilehvar. 2017. “On the Role of Text Preprocessing in Neural Network Architectures: An Evaluation Study on Text Categorization and Sentiment Analysis.” arXiv. https://doi.org/10.48550/ARXIV.1707.01780.
Cambria, Erik, Yang Li, Frank Z. Xing, Soujanya Poria, and Kenneth Kwok. 2020. “SenticNet 6: Ensemble Application of Symbolic and Subsymbolic AI for Sentiment Analysis.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412003.
Campos, David, Tung Kieu, Chenjuan Guo, Feiteng Huang, Kai Zheng, Bin Yang, and Christian S. Jensen. 2021. “Unsupervised Time Series Outlier Detection with Diversity-Driven Convolutional Ensembles.” Proceedings of the VLDB Endowment 15 (November). https://doi.org/10.14778/3494124.3494142.
Canchola, Jesse A. 2017. “Correct Use of Percent Coefficient of Variation (.” MOJ Proteomics &Amp; Bioinformatics 6 (November). https://doi.org/10.15406/mojpb.2017.06.00200.
Cañete, José, Gabriel Chaperon, Rodrigo Fuentes, Jou-Hui Ho, Hojin Kang, and Jorge Pérez. 2023. “Spanish Pre-Trained BERT Model and Evaluation Data.” arXiv. https://doi.org/10.48550/ARXIV.2308.02976.
Çano, Erion, and Ondřej Bojar. 2019. “Keyphrase Generation: A Text Summarization Struggle.” Proceedings of the 2019 Conference of the North. https://doi.org/10.18653/v1/n19-1070.
Cao, Bokai, Mia Mao, Siim Viidu, and Philip S. Yu. 2017. “HitFraud: A Broad Learning Approach for Collective Fraud Detection in Heterogeneous Information Networks.” arXiv. https://doi.org/10.48550/ARXIV.1709.04129.
Cao, Chen, Shihao Li, Shuo Yu, and Zhikui Chen. 2021. “Fake Reviewer Group Detection in Online Review Systems.” arXiv. https://doi.org/10.48550/ARXIV.2112.06403.
Cao, Jialun, Meiziniu Li, Ming Wen, and Shing-chi Cheung. 2023. “A Study on Prompt Design, Advantages and Limitations of ChatGPT for Deep Learning Program Repair.” arXiv. https://doi.org/10.48550/ARXIV.2304.08191.
Cao, Kaidi, Colin Wei, Adrien Gaidon, Nikos Arechiga, and Tengyu Ma. 2019. “Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss,” June. http://arxiv.org/abs/1906.07413v2.
Cao, Qingqing, Niranjan Balasubramanian, and Aruna Balasubramanian. 2017. “MobiRNN.” Proceedings of the 1st International Workshop on Deep Learning for Mobile Systems and Applications, June. https://doi.org/10.1145/3089801.3089804.
Cao, Shaosheng, Xinxing Yang, Cen Chen, Jun Zhou, Xiaolong Li, and Yuan Qi. 2019. “TitAnt: Online Real-Time Transaction Fraud Detection in Ant Financial.” arXiv. https://doi.org/10.48550/ARXIV.1906.07407.
Capper, David, David T. W. Jones, Martin Sill, Volker Hovestadt, Daniel Schrimpf, Dominik Sturm, Christian Koelsche, et al. 2018. “DNA Methylation-Based Classification of Central Nervous System Tumours.” Nature 555 (March). https://doi.org/10.1038/nature26000.
Caramalau, Razvan, Binod Bhattarai, and Tae-Kyun Kim. 2020. “Sequential Graph Convolutional Network for Active Learning.” arXiv. https://doi.org/10.48550/ARXIV.2006.10219.
Carballo, Kimberly Villalobos, Liangyuan Na, Yu Ma, Léonard Boussioux, Cynthia Zeng, Luis R. Soenksen, and Dimitris Bertsimas. 2022. “TabText: A Flexible and Contextual Approach to Tabular Data Representation.” arXiv. https://doi.org/10.48550/ARXIV.2206.10381.
Cardenas, Jaime, Carl B. Poitras, Jacob T. Robinson, Kyle Preston, Long Chen, and Michal Lipson. 2009. “Low Loss Etchless Silicon Photonic Waveguides.” Optics Express 17 (March). https://doi.org/10.1364/oe.17.004752.
Carlini, Nicholas, Anish Athalye, Nicolas Papernot, Wieland Brendel, Jonas Rauber, Dimitris Tsipras, Ian Goodfellow, Aleksander Madry, and Alexey Kurakin. 2019. “On Evaluating Adversarial Robustness.” arXiv. https://doi.org/10.48550/ARXIV.1902.06705.
Carlini, Nicholas, Matthew Jagielski, Christopher A. Choquette-Choo, Daniel Paleka, Will Pearce, Hyrum Anderson, Andreas Terzis, Kurt Thomas, and Florian Tramèr. 2023. “Poisoning Web-Scale Training Datasets Is Practical.” arXiv. https://doi.org/10.48550/ARXIV.2302.10149.
Carlini, Nicholas, Guy Katz, Clark Barrett, and David L. Dill. 2017. “Provably Minimally-Distorted Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1709.10207.
Carlini, Nicholas, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song. 2018. “The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1802.08232.
Carlini, Nicholas, and David Wagner. 2016. “Towards Evaluating the Robustness of Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1608.04644.
———. 2017. “MagNet and "Efficient Defenses Against Adversarial Attacks" Are Not Robust to Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1711.08478.
Carmel, David, Elad Haramaty, Arnon Lazerson, Liane Lewin-Eytan, and Yoelle Maarek. 2020. “Why Do People Buy Seemingly Irrelevant Items in Voice Product Search?” Proceedings of the 13th International Conference on Web Search and Data Mining, January. https://doi.org/10.1145/3336191.3371780.
Caron, Alberto, Gianluca Baio, and Ioanna Manolopoulou. 2022. “Estimating Individual Treatment Effects Using Non-Parametric Regression Models: A Review.” Journal of the Royal Statistical Society Series A: Statistics in Society 185 (March). https://doi.org/10.1111/rssa.12824.
Caron, Mathilde, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. 2021. “Emerging Properties in Self-Supervised Vision Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2104.14294.
Carreira-Perpiñán, Miguel Á., and Yerlan Idelbayev. 2017. “Model Compression as Constrained Optimization, with Application to Neural Nets. Part II: Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1707.04319.
Carreño, Ander, Iñaki Inza, and Jose A. Lozano. 2019. “Analyzing Rare Event, Anomaly, Novelty and Outlier Detection Terms Under the Supervised Classification Framework.” Artificial Intelligence Review 53 (October). https://doi.org/10.1007/s10462-019-09771-y.
Carta, Thomas, Clément Romac, Thomas Wolf, Sylvain Lamprier, Olivier Sigaud, and Pierre-Yves Oudeyer. 2023. “Grounding Large Language Models in Interactive Environments with Online Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.2302.02662.
Caruana, Rich, and Alexandru Niculescu-Mizil. 2006. “An Empirical Comparison of Supervised Learning Algorithms.” Proceedings of the 23rd International Conference on Machine Learning - ICML ’06. https://doi.org/10.1145/1143844.1143865.
Carvalho, Luiz M., Daniel A. M. Villela, Flavio C. Coelho, and Leonardo S. Bastos. 2023. “Bayesian Inference for the Weights in Logarithmic Pooling.” Bayesian Analysis 18 (March). https://doi.org/10.1214/22-ba1311.
Cassell, Justine, Hannes Högni Vilhjálmsson, and Timothy Bickmore. 2001. “BEAT.” Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, August. https://doi.org/10.1145/383259.383315.
Castellini, Claudio, and Patrick van der Smagt. 2008. “Surface EMG in Advanced Hand Prosthetics.” Biological Cybernetics 100 (November). https://doi.org/10.1007/s00422-008-0278-1.
Castro, Marcos R. de, Solange N. Gouvea, Andre Minella, Rafael Santos, and Nelson F. Souza-Sobrinho. 2015. “SAMBA: Stochastic Analytical Model with a Bayesian Approach.” Brazilian Review of Econometrics 35 (March). https://doi.org/10.12660/bre.v35n22015.57573.
Castro-Gonzalez, Leonardo, Yi-Ling Chung, Hannak Rose Kirk, John Francis, Angus R. Williams, Pica Johansson, and Jonathan Bright. 2024. “Cheap Learning: Maximising Performance of Language Models for Social Data Science Using Minimal Data.” arXiv. https://doi.org/10.48550/ARXIV.2401.12295.
Cavallari, Sandro, Vincent W. Zheng, Hongyun Cai, Kevin Chen-Chuan Chang, and Erik Cambria. 2017. “Learning Community Embedding with Community Detection and Node Embedding on Graphs.” Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3132847.3132925.
Cayton, Lawrence. 2008. “Fast Nearest Neighbor Retrieval for Bregman Divergences.” Proceedings of the 25th International Conference on Machine Learning - ICML ’08. https://doi.org/10.1145/1390156.1390171.
Cen, Yukuo, Jianwei Zhang, Xu Zou, Chang Zhou, Hongxia Yang, and Jie Tang. 2020. “Controllable Multi-Interest Framework for Recommendation.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403344.
Cen, Yukuo, Xu Zou, Jianwei Zhang, Hongxia Yang, Jingren Zhou, and Jie Tang. 2019. “Representation Learning for Attributed Multiplex Heterogeneous Network.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330964.
Cer, Daniel, Mona Diab, Eneko Agirre, Inigo Lopez-Gazpio, and Lucia Specia. 2017. “SemEval-2017 Task 1: Semantic Textual Similarity Multilingual and Crosslingual Focused Evaluation.” Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017). https://doi.org/10.18653/v1/s17-2001.
Cer, Daniel, Yinfei Yang, Sheng-yi Kong, Nan Hua, Nicole Limtiaco, Rhomni St. John, Noah Constant, et al. 2018. “Universal Sentence Encoder.” arXiv. https://doi.org/10.48550/ARXIV.1803.11175.
Češnovar, Rok, Steve Bronder, Davor Sluga, Jure Demšar, Tadej Ciglarič, Sean Talts, and Erik Štrumbelj. 2019. “GPU-Based Parallel Computation Support for Stan.” arXiv. https://doi.org/10.48550/ARXIV.1907.01063.
Chai, Yuning, Benjamin Sapp, Mayank Bansal, and Dragomir Anguelov. 2019. “MultiPath: Multiple Probabilistic Anchor Trajectory Hypotheses for Behavior Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1910.05449.
Chai, Ziwei, Yang Yang, Jiawang Dan, Sheng Tian, Changhua Meng, Weiqiang Wang, and Yifei Sun. 2023. “Towards Learning to Discover Money Laundering Sub-Network in Massive Transaction Network.” Proceedings of the AAAI Conference on Artificial Intelligence 37 (June). https://doi.org/10.1609/aaai.v37i12.26656.
Chai, Ziwei, Siqi You, Yang Yang, Shiliang Pu, Jiarong Xu, Haoyang Cai, and Weihao Jiang. 2022. “Can Abnormality Be Detected by Graph Neural Networks?” Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2022/270.
Chakravarti, Purvasha, Mikael Kuusela, Jing Lei, and Larry Wasserman. 2021. “Model-Independent Detection of New Physics Signals Using Interpretable Semi-Supervised Classifier Tests.” arXiv. https://doi.org/10.48550/ARXIV.2102.07679.
Chalkiadakis, Georgios, and Craig Boutilier. 2003. “Coordination in Multiagent Reinforcement Learning.” Proceedings of the Second International Joint Conference on Autonomous Agents and Multiagent Systems, July. https://doi.org/10.1145/860575.860689.
Champandard, Alex J. 2016. “Semantic Style Transfer and Turning Two-Bit Doodles into Fine Artworks.” arXiv. https://doi.org/10.48550/ARXIV.1603.01768.
Chan, Alvin, and Yew-Soon Ong. 2019. “Poison as a Cure: Detecting &Amp; Neutralizing Variable-Sized Backdoor Attacks in Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1911.08040.
Chan, Chunkit, Xin Liu, Jiayang Cheng, Zihan Li, Yangqiu Song, Ginny Y. Wong, and Simon See. 2023. “DiscoPrompt: Path Prediction Prompt Tuning for Implicit Discourse Relation Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2305.03973.
Chan, Stephanie C. Y., Adam Santoro, Andrew K. Lampinen, Jane X. Wang, Aaditya Singh, Pierre H. Richemond, Jay McClelland, and Felix Hill. 2022. “Data Distributional Properties Drive Emergent in-Context Learning in Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2205.05055.
Chan, William, Yu Zhang, Quoc Le, and Navdeep Jaitly. 2016. “Latent Sequence Decompositions.” arXiv. https://doi.org/10.48550/ARXIV.1610.03035.
Chandola, Varun, Arindam Banerjee, and Vipin Kumar. 2009. “Anomaly Detection.” ACM Computing Surveys 41 (July). https://doi.org/10.1145/1541880.1541882.
Chandra, Sulekh, and Poonam Pipil. 2013. “Spectral Studies of Transition Metal Complexes with 25, 26 Dioxo1,6,12,17,23,24 Hexaazacyclohexacosa 1,5,12,16 Tetraene Macrocyclic Ligand (l).” Open Journal of Inorganic Chemistry 03. https://doi.org/10.4236/ojic.2013.34013.
Chandrasekhar, Vijay, Jie Lin, Qianli Liao, Olivier Morère, Antoine Veillard, Lingyu Duan, and Tomaso Poggio. 2017. “Compression of Deep Neural Networks for Image Instance Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.1701.04923.
Chang, Angel X., Thomas Funkhouser, Leonidas Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio Savarese, et al. 2015. “ShapeNet: An Information-Rich 3D Model Repository,” December. http://arxiv.org/abs/1512.03012v1.
Chang, Chun-Hao, Elliot Creager, Anna Goldenberg, and David Duvenaud. 2018. “Explaining Image Classifiers by Counterfactual Generation.” arXiv. https://doi.org/10.48550/ARXIV.1807.08024.
Chang, Huiwen, Han Zhang, Jarred Barber, AJ Maschinot, Jose Lezama, Lu Jiang, Ming-Hsuan Yang, et al. 2023. “Muse: Text-to-Image Generation via Masked Generative Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2301.00704.
Chang, Hyun Sung, Yair Weiss, and William T. Freeman. 2009. “Informative Sensing.” arXiv. https://doi.org/10.48550/ARXIV.0901.4275.
Chang, Kai-Wei, Wei-Cheng Tseng, Shang-Wen Li, and Hung-yi Lee. 2022. “SpeechPrompt: An Exploration of Prompt Tuning on Generative Spoken Language Model for Speech Processing Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2203.16773.
Chang, Minsuk, Stefania Druga, Alex Fiannaca, Pedro Vergani, Chinmay Kulkarni, Carrie Cai, and Michael Terry. 2023. “The Prompt Artists,” March. http://arxiv.org/abs/2303.12253v1.
Chang, Yupeng, Xu Wang, Jindong Wang, Yuan Wu, Linyi Yang, Kaijie Zhu, Hao Chen, et al. 2023. “A Survey on Evaluation of Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2307.03109.
Changpinyo, Soravit, Mark Sandler, and Andrey Zhmoginov. 2017. “The Power of Sparsity in Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1702.06257.
Chapelle, Olivier. 2014. “Modeling Delayed Feedback in Display Advertising.” Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2623330.2623634.
———. 2015. “Offline Evaluation of Response Prediction in Online Advertising Auctions.” Proceedings of the 24th International Conference on World Wide Web, May. https://doi.org/10.1145/2740908.2742566.
Chapelle, Olivier, Eren Manavoglu, and Romer Rosales. 2014. “Simple and Scalable Response Prediction for Display Advertising.” ACM Transactions on Intelligent Systems and Technology 5 (December). https://doi.org/10.1145/2532128.
Charte, Francisco, and David Charte. 2015. “Working with Multilabel Datasets in r: The Mldr Package.” The R Journal 7. https://doi.org/10.32614/rj-2015-027.
Chassaing, Philippe, and Gilles Schaeffer. 2003. “Random Planar Lattices and Integrated superBrownian Excursion.” Probability Theory and Related Fields 128 (November). https://doi.org/10.1007/s00440-003-0297-8.
Chatfield, Ken, Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. 2014. “Return of the Devil in the Details: Delving Deep into Convolutional Nets.” arXiv. https://doi.org/10.48550/ARXIV.1405.3531.
“ChatGPT for Research and Publication: Opportunities and Challenges.” 2023. 1 6 (April). https://doi.org/10.37074/jalt.2023.6.1.20.
Chatterjee, Joyjit, and Nina Dethlefs. 2023. “This New Conversational AI Model Can Be Your Friend, Philosopher, and Guide ... And Even Your Worst Enemy.” Patterns 4 (January). https://doi.org/10.1016/j.patter.2022.100676.
Chatterjee, Rajdeep, Saptarshi Mazumdar, R. Simon Sherratt, Rohit Halder, Tanmoy Maitra, and Debasis Giri. 2021. “Real-Time Speech Emotion Analysis for Smart Home Assistants.” IEEE Transactions on Consumer Electronics 67 (February). https://doi.org/10.1109/tce.2021.3056421.
Chatterjee, Sourav. 2007. “Fluctuations of Eigenvalues and Second Order Poincaré Inequalities.” Probability Theory and Related Fields 143 (November). https://doi.org/10.1007/s00440-007-0118-6.
Chaudhuri, Siddhartha, Daniel Ritchie, and Kai Xu. 2019. “Learning Generative Models of 3D Structures.” Eurographics 2019 - Tutorials. https://doi.org/10.2312/EGT.20191038.
Chauvin, Brigitte, and Alain Rouault. 1988. “KPP Equation and Supercritical Branching Brownian Motion in the Subcritical Speed Area. Application to Spatial Trees.” Probability Theory and Related Fields 80 (December). https://doi.org/10.1007/bf00356108.
Chavent, Marie, Vanessa Kuentz-Simonet, Amaury Labenne, and Jérôme Saracco. 2018. “ClustGeo: An r Package for Hierarchical Clustering with Spatial Constraints.” Computational Statistics 33 (January). https://doi.org/10.1007/s00180-018-0791-1.
Che, Fengdi, Xiru Zhu, Doina Precup, David Meger, and Gregory Dudek. 2022. “Bayesian q-Learning with Imperfect Expert Demonstrations.” arXiv. https://doi.org/10.48550/ARXIV.2210.01800.
Che, Tong, Yanran Li, Ruixiang Zhang, R Devon Hjelm, Wenjie Li, Yangqiu Song, and Yoshua Bengio. 2017. “Maximum-Likelihood Augmented Discrete Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1702.07983.
Che, Zhengping, Sanjay Purushotham, Kyunghyun Cho, David Sontag, and Yan Liu. 2016. “Recurrent Neural Networks for Multivariate Time Series with Missing Values.” arXiv. https://doi.org/10.48550/ARXIV.1606.01865.
Chekalina, Viktoriia, and Alexander Panchenko. 2023. “Retrieving Comparative Arguments Using Ensemble Methods and Neural Information Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.2305.01513.
Chelba, Ciprian, Tomas Mikolov, Mike Schuster, Qi Ge, Thorsten Brants, Phillipp Koehn, and Tony Robinson. 2013. “One Billion Word Benchmark for Measuring Progress in Statistical Language Modeling.” arXiv. https://doi.org/10.48550/ARXIV.1312.3005.
Chelombiev, Ivan, Conor Houghton, and Cian O’Donnell. 2019. “Adaptive Estimators Show Information Compression in Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1902.09037.
Chen, Chaoqi, Weiping Xie, Wenbing Huang, Yu Rong, Xinghao Ding, Yue Huang, Tingyang Xu, and Junzhou Huang. 2018. “Progressive Feature Alignment for Unsupervised Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1811.08585.
Chen, Charlie, Sebastian Borgeaud, Geoffrey Irving, Jean-Baptiste Lespiau, Laurent Sifre, and John Jumper. 2023. “Accelerating Large Language Model Decoding with Speculative Sampling.” arXiv. https://doi.org/10.48550/ARXIV.2302.01318.
Chen, Danqi, Jason Bolton, and Christopher D. Manning. 2016. “A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task.” arXiv. https://doi.org/10.48550/ARXIV.1606.02858.
Chen, Daoyuan, Yaliang Li, Min Yang, Hai-Tao Zheng, and Ying Shen. 2019. “Knowledge-Aware Textual Entailment with Graph Attention Network.” Proceedings of the 28th ACM International Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3357384.3358071.
Chen, Dave Zhenyu, Qirui Wu, Matthias Nießner, and Angel X. Chang. 2021. “D3Net: A Unified Speaker-Listener Architecture for 3D Dense Captioning and Visual Grounding.” arXiv. https://doi.org/10.48550/ARXIV.2112.01551.
Chen, Guangyao, Peixi Peng, Xiangqian Wang, and Yonghong Tian. 2021. “Adversarial Reciprocal Points Learning for Open Set Recognition.” IEEE Transactions on Pattern Analysis and Machine Intelligence. https://doi.org/10.1109/tpami.2021.3106743.
Chen, Guangyi, Weiran Yao, Xiangchen Song, Xinyue Li, Yongming Rao, and Kun Zhang. 2022. “PLOT: Prompt Learning with Optimal Transport for Vision-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2210.01253.
Chen, Haochen, Bryan Perozzi, Yifan Hu, and Steven Skiena. 2017. “HARP: Hierarchical Representation Learning for Networks.” arXiv. https://doi.org/10.48550/ARXIV.1706.07845.
Chen, Haochen, Xiaofei Sun, Yingtao Tian, Bryan Perozzi, Muhao Chen, and Steven Skiena. 2018. “Enhanced Network Embeddings via Exploiting Edge Labels.” arXiv. https://doi.org/10.48550/ARXIV.1809.05124.
Chen, Irene Y., Emma Pierson, Sherri Rose, Shalmali Joshi, Kadija Ferryman, and Marzyeh Ghassemi. 2021. “Ethical Machine Learning in Healthcare.” Annual Review of Biomedical Data Science 4 (July). https://doi.org/10.1146/annurev-biodatasci-092820-114757.
Chen, J., and Z. Chen. 2008. “Extended Bayesian Information Criteria for Model Selection with Large Model Spaces.” Biometrika 95 (September). https://doi.org/10.1093/biomet/asn034.
Chen, Jianfei, Jun Zhu, and Le Song. 2017. “Stochastic Training of Graph Convolutional Networks with Variance Reduction.” arXiv. https://doi.org/10.48550/ARXIV.1710.10568.
Chen, Jia, Peng Wang, and Wei Wang. 2022. “Online Summarizing Alerts Through Semantic and Behavior Information.” Proceedings of the 44th International Conference on Software Engineering, May. https://doi.org/10.1145/3510003.3510055.
Chen, Jing, and Silvio Micali. 2016. “Algorand.” arXiv. https://doi.org/10.48550/ARXIV.1607.01341.
Chen, Jinyin, Yixian Chen, Lihong Chen, Minghao Zhao, and Qi Xuan. 2019. “Multiscale Evolutionary Perturbation Attack on Community Detection.” arXiv. https://doi.org/10.48550/ARXIV.1910.09741.
Chen, Jinyin, Haiyang Xiong, Dunjie Zhang, Zhenguang Liu, and Jiajing Wu. 2021. “TEGDetector: A Phishing Detector That Knows Evolving Transaction Behaviors.” arXiv. https://doi.org/10.48550/ARXIV.2111.15446.
Chen, Kai, Mathias Seuret, Marcus Liwicki, Jean Hennebert, and Rolf Ingold. 2015. “Page Segmentation of Historical Document Images with Convolutional Autoencoders.” 2015 13th International Conference on Document Analysis and Recognition (ICDAR), August. https://doi.org/10.1109/icdar.2015.7333914.
Chen, Kai, Jiaqi Wang, Jiangmiao Pang, Yuhang Cao, Yu Xiong, Xiaoxiao Li, Shuyang Sun, et al. 2019. “MMDetection: Open MMLab Detection Toolbox and Benchmark.” arXiv. https://doi.org/10.48550/ARXIV.1906.07155.
Chen, Ken, Fei Chen, Baisheng Lai, Zhongming Jin, Yong Liu, Kai Li, Long Wei, et al. 2018. “Dynamic Spatio-Temporal Graph-Based CNNs for Traffic Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1812.02019.
Chen, Ke, Yaoqi Zhou, Sheng Wang, and Peng Xiong. 2023. “<Scp>RNA</Scp> Tertiary Structure Modeling with <Scp>BRiQ</Scp> Potential in <Scp>CASP15</Scp>.” Proteins: Structure, Function, and Bioinformatics 91 (August). https://doi.org/10.1002/prot.26574.
Chen, Lele, Sudhanshu Srivastava, Zhiyao Duan, and Chenliang Xu. 2017. “Deep Cross-Modal Audio-Visual Generation.” arXiv. https://doi.org/10.48550/ARXIV.1704.08292.
Chen, Liang-Chieh, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L. Yuille. 2014. “Semantic Image Segmentation with Deep Convolutional Nets and Fully Connected CRFs.” arXiv. https://doi.org/10.48550/ARXIV.1412.7062.
———. 2016. “DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.” arXiv. https://doi.org/10.48550/ARXIV.1606.00915.
Chen, Liang-Chieh, George Papandreou, Florian Schroff, and Hartwig Adam. 2017. “Rethinking Atrous Convolution for Semantic Image Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1706.05587.
Chen, Li-Wei, Shinji Watanabe, and Alexander Rudnicky. 2023. “A Vector Quantized Approach for Text to Speech Synthesis on Real-World Spontaneous Speech.” arXiv. https://doi.org/10.48550/ARXIV.2302.04215.
Chen, Mia Xu, Benjamin N Lee, Gagan Bansal, Yuan Cao, Shuyuan Zhang, Justin Lu, Jackie Tsay, et al. 2019. “Gmail Smart Compose: Real-Time Assisted Writing.” arXiv. https://doi.org/10.48550/ARXIV.1906.00080.
Chen, Ming-Hui, and Qi-Man Shao. 1999. “Monte Carlo Estimation of Bayesian Credible and HPD Intervals.” Journal of Computational and Graphical Statistics 8 (March). https://doi.org/10.1080/10618600.1999.10474802.
Chen, Mingjian, Xu Tan, Bohan Li, Yanqing Liu, Tao Qin, Sheng Zhao, and Tie-Yan Liu. 2021. “AdaSpeech: Adaptive Text to Speech for Custom Voice.” arXiv. https://doi.org/10.48550/ARXIV.2103.00993.
Chen, Muhao, Yingtao Tian, Mohan Yang, and Carlo Zaniolo. 2016. “Multilingual Knowledge Graph Embeddings for Cross-Lingual Knowledge Alignment.” arXiv. https://doi.org/10.48550/ARXIV.1611.03954.
Chen, Qiwei, Huan Zhao, Wei Li, Pipei Huang, and Wenwu Ou. 2019. “Behavior Sequence Transformer for e-Commerce Recommendation in Alibaba.” arXiv. https://doi.org/10.48550/ARXIV.1905.06874.
Chen, Shizhe, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, and Ivan Laptev. 2022. “Language Conditioned Spatial Relation Reasoning for 3D Object Grounding.” arXiv. https://doi.org/10.48550/ARXIV.2211.09646.
Chen, Shu-Yu, Jia-Qi Zhang, You-You Zhao, Paul L. Rosin, Yu-Kun Lai, and Lin Gao. 2022. “A Review of Image and Video Colorization: From Analogies to Deep Learning.” Visual Informatics 6 (September). https://doi.org/10.1016/j.visinf.2022.05.003.
Chen, Sijia, Bin Song, Xiaojiang Du, and Nadra Guizani. 2019. “Structured Bayesian Compression for Deep Models in Mobile Enabled Devices for Connected Healthcare.” arXiv. https://doi.org/10.48550/ARXIV.1902.05429.
Chen, Tianqi, Ian Goodfellow, and Jonathon Shlens. 2015. “Net2Net: Accelerating Learning via Knowledge Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1511.05641.
Chen, Tianqi, Mu Li, Yutian Li, Min Lin, Naiyan Wang, Minjie Wang, Tianjun Xiao, Bing Xu, Chiyuan Zhang, and Zheng Zhang. 2015. “MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous Distributed Systems.” arXiv. https://doi.org/10.48550/ARXIV.1512.01274.
Chen, Tianqi, Bing Xu, Chiyuan Zhang, and Carlos Guestrin. 2016. “Training Deep Nets with Sublinear Memory Cost.” arXiv. https://doi.org/10.48550/ARXIV.1604.06174.
Chen, Tianqi, Zhao Zheng, Qiuxia Lu, Weinan Zhang, and Yong Yu. 2011. “Feature-Based Matrix Factorization.” arXiv. https://doi.org/10.48550/ARXIV.1109.2271.
Chen, Tianshi, Zidong Du, Ninghui Sun, Jia Wang, Chengyong Wu, Yunji Chen, and Olivier Temam. 2014. “DianNao.” Proceedings of the 19th International Conference on Architectural Support for Programming Languages and Operating Systems, February. https://doi.org/10.1145/2541940.2541967.
Chen, Tianyi, and Charalampos E. Tsourakakis. 2022. “AntiBenford Subgraphs: Unsupervised Anomaly Detection in Financial Networks.” arXiv. https://doi.org/10.48550/ARXIV.2205.13426.
Chen, Timothy, Preston Culbertson, and Mac Schwager. 2023. “CATNIPS: Collision Avoidance Through Neural Implicit Probabilistic Scenes.” arXiv. https://doi.org/10.48550/ARXIV.2302.12931.
Chen, Ting, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. 2020. “A Simple Framework for Contrastive Learning of Visual Representations.” arXiv. https://doi.org/10.48550/ARXIV.2002.05709.
Chen, Ting, and Yizhou Sun. 2017. “Task-Guided and Path-Augmented Heterogeneous Network Embedding for Author Identification.” Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/3018661.3018735.
Chen, Wei, Jincai Chen, Fuhao Zou, Yuan-Fang Li, Ping Lu, and Wei Zhao. 2019. “RobustiQ.” Proceedings of the 2019 on International Conference on Multimedia Retrieval, June. https://doi.org/10.1145/3323873.3325018.
Chen, Wei-Chun, Chia-Che Chang, Chien-Yu Lu, and Che-Rung Lee. 2018. “Knowledge Distillation with Feature Maps for Image Classification.” arXiv. https://doi.org/10.48550/ARXIV.1812.00660.
Chen, Weijie, Yuan Zhang, Di Xie, and Shiliang Pu. 2018. “A Layer Decomposition-Recomposition Framework for Neuron Pruning Towards Accurate Lightweight Networks.” arXiv. https://doi.org/10.48550/ARXIV.1812.06611.
Chen, Weikai, Cheng Lin, Weiyang Li, and Bo Yang. 2022. “3PSDF: Three-Pole Signed Distance Function for Learning Surfaces with Arbitrary Topologies.” arXiv. https://doi.org/10.48550/ARXIV.2205.15572.
Chen, Wei-Yu, Yen-Cheng Liu, Zsolt Kira, Yu-Chiang Frank Wang, and Jia-Bin Huang. 2019. “A Closer Look at Few-Shot Classification.” arXiv. https://doi.org/10.48550/ARXIV.1904.04232.
Chen, Wenhu, Jianshu Chen, Pengda Qin, Xifeng Yan, and William Yang Wang. 2019. “Semantically Conditioned Dialog Response Generation via Hierarchical Disentangled Self-Attention.” arXiv. https://doi.org/10.48550/ARXIV.1905.12866.
Chen, Wenhu, Hongmin Wang, Jianshu Chen, Yunkai Zhang, Hong Wang, Shiyang Li, Xiyou Zhou, and William Yang Wang. 2019. “TabFact: A Large-Scale Dataset for Table-Based Fact Verification.” arXiv. https://doi.org/10.48550/ARXIV.1909.02164.
Chen, Wenlin, James T. Wilson, Stephen Tyree, Kilian Q. Weinberger, and Yixin Chen. 2015. “Compressing Neural Networks with the Hashing Trick.” arXiv. https://doi.org/10.48550/ARXIV.1504.04788.
Chen, Xiangning, Cho-Jui Hsieh, and Boqing Gong. 2021. “When Vision Transformers Outperform ResNets Without Pre-Training or Strong Data Augmentations.” arXiv. https://doi.org/10.48550/ARXIV.2106.01548.
Chen, Xiangning, Chen Liang, Da Huang, Esteban Real, Kaiyuan Wang, Yao Liu, Hieu Pham, et al. 2023. “Symbolic Discovery of Optimization Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.2302.06675.
Chen, Xia, Guoxian Yu, Jun Wang, Carlotta Domeniconi, Zhao Li, and Xiangliang Zhang. 2019. “ActiveHNE: Active Heterogeneous Network Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1905.05659.
Chen, Xilun, and Claire Cardie. 2018. “Unsupervised Multilingual Word Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1808.08933.
Chen, Xingyu, Zihan Zhao, Lu Chen, Danyang Zhang, Jiabao Ji, Ao Luo, Yuxuan Xiong, and Kai Yu. 2021. “WebSRC: A Dataset for Web-Based Structural Reading Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.2101.09465.
Chen, Xinlei, Haoqi Fan, Ross Girshick, and Kaiming He. 2020. “Improved Baselines with Momentum Contrastive Learning.” arXiv. https://doi.org/10.48550/ARXIV.2003.04297.
Chen, Xinlei, Hao Fang, Tsung-Yi Lin, Ramakrishna Vedantam, Saurabh Gupta, Piotr Dollar, and C. Lawrence Zitnick. 2015. “Microsoft COCO Captions: Data Collection and Evaluation Server.” arXiv. https://doi.org/10.48550/ARXIV.1504.00325.
Chen, Xinlei, and C. Lawrence Zitnick. 2014. “Learning a Recurrent Visual Representation for Image Caption Generation.” arXiv. https://doi.org/10.48550/ARXIV.1411.5654.
Chen, Xinyun, Maxwell Lin, Nathanael Schärli, and Denny Zhou. 2023. “Teaching Large Language Models to Self-Debug.” arXiv. https://doi.org/10.48550/ARXIV.2304.05128.
Chen, Xinyun, Chang Liu, Bo Li, Kimberly Lu, and Dawn Song. 2017. “Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning.” arXiv. https://doi.org/10.48550/ARXIV.1712.05526.
Chen, Xumin, Peng Cui, Lingling Yi, and Shiqiang Yang. 2018. “Scalable Optimization for Embedding Highly-Dynamic and Recency-Sensitive Data.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3219898.
Chen, Yen-Chi, Daren Wang, Alessandro Rinaldo, and Larry Wasserman. 2015. “Statistical Analysis of Persistence Intensity Functions.” arXiv. https://doi.org/10.48550/ARXIV.1510.02502.
Chen, Yinpeng, Xiyang Dai, Dongdong Chen, Mengchen Liu, Xiaoyi Dong, Lu Yuan, and Zicheng Liu. 2021. “Mobile-Former: Bridging MobileNet and Transformer.” arXiv. https://doi.org/10.48550/ARXIV.2108.05895.
Chen, Yubo, Shulin Liu, Xiang Zhang, Kang Liu, and Jun Zhao. 2017. “Automatically Labeled Data Generation for Large Scale Event Extraction.” Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/p17-1038.
Chen, Yu-Hsin, Tushar Krishna, Joel S. Emer, and Vivienne Sze. 2017. “Eyeriss: An Energy-Efficient Reconfigurable Accelerator for Deep Convolutional Neural Networks.” IEEE Journal of Solid-State Circuits 52 (January). https://doi.org/10.1109/jssc.2016.2616357.
Chen, Yukang, Gaofeng Meng, Qian Zhang, Xinbang Zhang, Liangchen Song, Shiming Xiang, and Chunhong Pan. 2018. “Joint Neural Architecture Search and Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1811.09426.
Chen, Yukang, Tong Yang, Xiangyu Zhang, Gaofeng Meng, Xinyu Xiao, and Jian Sun. 2019. “DetNAS: Backbone Search for Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.1903.10979.
Chen, Yulong, Yang Liu, Li Dong, Shuohang Wang, Chenguang Zhu, Michael Zeng, and Yue Zhang. 2022. “AdaPrompt: Adaptive Model Training for Prompt-Based NLP.” arXiv. https://doi.org/10.48550/ARXIV.2202.04824.
Chen, Yunpeng, Jianshu Li, Bin Zhou, Jiashi Feng, and Shuicheng Yan. 2017. “Weaving Multi-Scale Context for Single Shot Detector.” arXiv. https://doi.org/10.48550/ARXIV.1712.03149.
Chen, Yuntao, Naiyan Wang, and Zhaoxiang Zhang. 2017. “DarkRank: Accelerating Deep Metric Learning via Cross Sample Similarities Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1707.01220.
Chen, Yu, Lingfei Wu, and Mohammed J. Zaki. 2024. “Toward Subgraph-Guided Knowledge Graph Question Generation with Graph Neural Networks.” IEEE Transactions on Neural Networks and Learning Systems. https://doi.org/10.1109/tnnls.2023.3264519.
Chen, Zhehuai, Ankur Bapna, Andrew Rosenberg, Yu Zhang, Bhuvana Ramabhadran, Pedro Moreno, and Nanxin Chen. 2022. “Maestro-u: Leveraging Joint Speech-Text Representation Learning for Zero Supervised Speech ASR.” arXiv. https://doi.org/10.48550/ARXIV.2210.10027.
Chen, Zhenpeng, Sheng Shen, Ziniu Hu, Xuan Lu, Qiaozhu Mei, and Xuanzhe Liu. 2019. “Emoji-Powered Representation Learning for Cross-Lingual Sentiment Classification.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313600.
Chen, Zhen-Qing, and Takashi Kumagai. 2007. “Heat Kernel Estimates for Jump Processes of Mixed Types on Metric Measure Spaces.” Probability Theory and Related Fields 140 (April). https://doi.org/10.1007/s00440-007-0070-5.
Chen, Zhuo, Weisi Lin, Shiqi Wang, Lingyu Duan, and Alex C. Kot. 2018. “Intermediate Deep Feature Compression: The Next Battlefield of Intelligent Sensing.” arXiv. https://doi.org/10.48550/ARXIV.1809.06196.
Chen, Zhuo, Jiyuan Zhang, Ruizhou Ding, and Diana Marculescu. 2019. “ViP: Virtual Pooling for Accelerating CNN-Based Image Classification and Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.1906.07912.
Cheng, Chen, Fen Xia, Tong Zhang, Irwin King, and Michael R. Lyu. 2014. “Gradient Boosting Factorization Machines.” Proceedings of the 8th ACM Conference on Recommender Systems, October. https://doi.org/10.1145/2645710.2645730.
Cheng, Heng-Tze, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, et al. 2016. “Wide &Amp; Deep Learning for Recommender Systems.” arXiv. https://doi.org/10.48550/ARXIV.1606.07792.
Cheng, Jianpeng, Li Dong, and Mirella Lapata. 2016. “Long Short-Term Memory-Networks for Machine Reading.” Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/d16-1053.
Cheng, Jian, Pei-song Wang, Gang Li, Qing-hao Hu, and Han-qing Lu. 2018. “Recent Advances in Efficient Computation of Deep Convolutional Neural Networks.” Frontiers of Information Technology &Amp; Electronic Engineering 19 (January). https://doi.org/10.1631/fitee.1700789.
Cheng, Jun, Guido Novati, Joshua Pan, Clare Bycroft, Akvilė Žemgulytė, Taylor Applebaum, Alexander Pritzel, et al. 2023. “Accurate Proteome-Wide Missense Variant Effect Prediction with AlphaMissense.” Science 381 (September). https://doi.org/10.1126/science.adg7492.
Cheng, Taoli, and Aaron Courville. 2023. “Versatile Energy-Based Probabilistic Models for High Energy Physics.” arXiv. https://doi.org/10.48550/ARXIV.2302.00695.
Cheng, Tiejun, Yuan Zhao, Xun Li, Fu Lin, Yong Xu, Xinglong Zhang, Yan Li, Renxiao Wang, and Luhua Lai. 2007. “Computation of Octanol−water Partition Coefficients by Guiding an Additive Model with Knowledge.” Journal of Chemical Information and Modeling 47 (November). https://doi.org/10.1021/ci700257y.
Cheng, Wei, Kai Zhang, Haifeng Chen, Guofei Jiang, Zhengzhang Chen, and Wei Wang. 2016. “Ranking Causal Anomalies via Temporal and Dynamical Analysis on Vanishing Correlations.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939765.
Cheng, Xi, Bohdan Khomtchouk, Norman Matloff, and Pete Mohanty. 2018. “Polynomial Regression as an Alternative to Neural Nets.” arXiv. https://doi.org/10.48550/ARXIV.1806.06850.
Cheng, Xueqi, Yixing Fan, Jiafeng Guo, Yiyi Liu, and Ruqing Zhang. 2023. “Prompt Tuning with Contradictory Intentions for Sarcasm Recognition.” https://doi.org/10.48448/BQXW-B671.
Cheng, Yu, Duo Wang, Pan Zhou, and Tao Zhang. 2017. “A Survey of Model Compression and Acceleration for Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1710.09282.
———. 2018. “Model Compression and Acceleration for Deep Neural Networks: The Principles, Progress, and Challenges.” IEEE Signal Processing Magazine 35 (January). https://doi.org/10.1109/msp.2017.2765695.
Cheng, Yu, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, and Shi-Fu Chang. 2015. “An Exploration of Parameter Redundancy in Deep Networks with Circulant Projections.” 2015 IEEE International Conference on Computer Vision (ICCV), December. https://doi.org/10.1109/iccv.2015.327.
Cheng, Zehua, and Zhenghua Xu. 2019. “Bandwidth Reduction Using Importance Weighted Pruning on Ring AllReduce.” arXiv. https://doi.org/10.48550/ARXIV.1901.01544.
Cheng, Zhiyong, Daniel Soudry, Zexi Mao, and Zhenzhong Lan. 2015. “Training Binary Multilayer Neural Networks for Image Classification Using Expectation Backpropagation.” arXiv. https://doi.org/10.48550/ARXIV.1503.03562.
Cheraghian, Ali, Shafin Rahman, Dylan Campbell, and Lars Petersson. 2019. “Transductive Zero-Shot Learning for 3D Point Cloud Classification.” arXiv. https://doi.org/10.48550/ARXIV.1912.07161.
Cherepanova, Valeriia, Micah Goldblum, Harrison Foley, Shiyuan Duan, John Dickerson, Gavin Taylor, and Tom Goldstein. 2021. “LowKey: Leveraging Adversarial Attacks to Protect Social Media Users from Facial Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2101.07922.
Chernozhukov, Victor, Denis Chetverikov, and Kengo Kato. 2014. “Comparison and Anti-Concentration Bounds for Maxima of Gaussian Random Vectors.” Probability Theory and Related Fields 162 (May). https://doi.org/10.1007/s00440-014-0565-9.
Chetlur, Sharan, Cliff Woolley, Philippe Vandermersch, Jonathan Cohen, John Tran, Bryan Catanzaro, and Evan Shelhamer. 2014. “cuDNN: Efficient Primitives for Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1410.0759.
Chetrite, Raphaël, and Shamik Gupta. 2011. “Two Refreshing Views of Fluctuation Theorems Through Kinematics Elements and Exponential Martingale.” Journal of Statistical Physics 143 (April). https://doi.org/10.1007/s10955-011-0184-0.
Chi, Zewen, Li Dong, Furu Wei, Wenhui Wang, Xian-Ling Mao, and Heyan Huang. 2019. “Cross-Lingual Natural Language Generation via Pre-Training.” arXiv. https://doi.org/10.48550/ARXIV.1909.10481.
Chickering, David Maxwell, David Heckerman, and Christopher Meek. 2013. “A Bayesian Approach to Learning Bayesian Networks with Local Structure.” arXiv. https://doi.org/10.48550/ARXIV.1302.1528.
Chierichetti, Flavio, Alessandro Epasto, Ravi Kumar, Silvio Lattanzi, and Vahab Mirrokni. 2015. “Efficient Algorithms for Public-Private Social Networks.” Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2783258.2783354.
Child, Rewon, Scott Gray, Alec Radford, and Ilya Sutskever. 2019. “Generating Long Sequences with Sparse Transformers.” arXiv. https://doi.org/10.48550/ARXIV.1904.10509.
Chipman, Hugh A., Edward I. George, and Robert E. McCulloch. 2007. “Bayesian Ensemble Learning.” Advances in Neural Information Processing Systems 19, September. https://doi.org/10.7551/mitpress/7503.003.0038.
Chipman, Hugh A., Edward I. George, Robert E. McCulloch, and Thomas S. Shively. 2022. “mBART: Multidimensional Monotone BART.” Bayesian Analysis 17 (June). https://doi.org/10.1214/21-ba1259.
Chiquet, Julien, Mahendra Mariadassou, and Stéphane Robin. 2017. “Variational Inference for Probabilistic Poisson PCA.” arXiv. https://doi.org/10.48550/ARXIV.1703.06633.
Chiu, Chih-Yao, Hwann-Tzong Chen, and Tyng-Luh Liu. 2019. “C2S2: Cost-Aware Channel Sparse Selection for Progressive Network Pruning.” arXiv. https://doi.org/10.48550/ARXIV.1904.03508.
Chiu, Chung-Cheng, Dieterich Lawson, Yuping Luo, George Tucker, Kevin Swersky, Ilya Sutskever, and Navdeep Jaitly. 2017. “An Online Sequence-to-Sequence Model for Noisy Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1706.06428.
Chiu, Chung-Cheng, and Colin Raffel. 2017. “Monotonic Chunkwise Attention.” arXiv. https://doi.org/10.48550/ARXIV.1712.05382.
Chiu, Chung-Cheng, Tara N. Sainath, Yonghui Wu, Rohit Prabhavalkar, Patrick Nguyen, Zhifeng Chen, Anjuli Kannan, et al. 2017. “State-of-the-Art Speech Recognition with Sequence-to-Sequence Models.” arXiv. https://doi.org/10.48550/ARXIV.1712.01769.
Chizat, Lenaic, Gabriel Peyré, Bernhard Schmitzer, and François-Xavier Vialard. 2016. “Scaling Algorithms for Unbalanced Transport Problems.” arXiv. https://doi.org/10.48550/ARXIV.1607.05816.
Cho, Hyeoncheol, Youngrock Oh, and Eunjoo Jeon. 2021. “SEEN: Sharpening Explanations for Graph Neural Networks Using Explanations from Neighborhoods.” arXiv. https://doi.org/10.48550/ARXIV.2106.08532.
Cho, Hyunsoo, Jinseok Seol, and Sang-goo Lee. 2021. “Masked Contrastive Learning for Anomaly Detection.” arXiv. https://doi.org/10.48550/ARXIV.2105.08793.
Cho, Kyunghyun, Bart van Merrienboer, Dzmitry Bahdanau, and Yoshua Bengio. 2014. “On the Properties of Neural Machine Translation: Encoder-Decoder Approaches.” arXiv. https://doi.org/10.48550/ARXIV.1409.1259.
Cho, Kyunghyun, Bart van Merrienboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. “Learning Phrase Representations Using RNN Encoder–Decoder for Statistical Machine Translation.” Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). https://doi.org/10.3115/v1/d14-1179.
Choi, Byoung Jin, Myeonghun Jeong, Minchan Kim, Sung Hwan Mun, and Nam Soo Kim. 2022. “Adversarial Speaker-Consistency Learning Using Untranscribed Speech Data for Zero-Shot Multi-Speaker Text-to-Speech.” arXiv. https://doi.org/10.48550/ARXIV.2210.05979.
Choi, Edward, Mohammad Taha Bahadori, Andy Schuetz, Walter F. Stewart, and Jimeng Sun. 2015. “Doctor AI: Predicting Clinical Events via Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.05942.
Choi, Edward, Andy Schuetz, Walter F Stewart, and Jimeng Sun. 2016. “Using Recurrent Neural Network Models for Early Detection of Heart Failure Onset.” Journal of the American Medical Informatics Association 24 (August). https://doi.org/10.1093/jamia/ocw112.
Choi, Myung Jin, Vincent Y. F. Tan, Animashree Anandkumar, and Alan S. Willsky. 2010. “Learning Latent Tree Graphical Models.” arXiv. https://doi.org/10.48550/ARXIV.1009.2722.
Choi, Yoojin, Mostafa El-Khamy, and Jungwon Lee. 2016. “Towards the Limit of Network Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1612.01543.
———. 2020. “Learning Sparse Low-Precision Neural Networks with Learnable Regularization.” IEEE Access 8. https://doi.org/10.1109/access.2020.2996936.
Chollet, Francois. 2017. “Xception: Deep Learning with Depthwise Separable Convolutions.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July. https://doi.org/10.1109/cvpr.2017.195.
Chollet, François. 2016. “Xception: Deep Learning with Depthwise Separable Convolutions.” arXiv. https://doi.org/10.48550/ARXIV.1610.02357.
Chong, Penny, Lukas Ruff, Marius Kloft, and Alexander Binder. 2020. “Simple and Effective Prevention of Mode Collapse in Deep One-Class Classification.” 2020 International Joint Conference on Neural Networks (IJCNN), July. https://doi.org/10.1109/ijcnn48605.2020.9207209.
Choromanska, Anna, Mikael Henaff, Michael Mathieu, Gérard Ben Arous, and Yann LeCun. 2014. “The Loss Surfaces of Multilayer Networks.” arXiv. https://doi.org/10.48550/ARXIV.1412.0233.
Chorowski, Jan, Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. “End-to-End Continuous Speech Recognition Using Attention-Based Recurrent NN: First Results.” arXiv. https://doi.org/10.48550/ARXIV.1412.1602.
Chorowski, Jan, and Navdeep Jaitly. 2016. “Towards Better Decoding and Language Model Integration in Sequence to Sequence Models.” arXiv. https://doi.org/10.48550/ARXIV.1612.02695.
———. 2017. “Towards Better Decoding and Language Model Integration in Sequence to Sequence Models.” Interspeech 2017, August. https://doi.org/10.21437/interspeech.2017-343.
Chow, C., and C. Liu. 1968. “Approximating Discrete Probability Distributions with Dependence Trees.” IEEE Transactions on Information Theory 14 (May). https://doi.org/10.1109/tit.1968.1054142.
Chowdhery, Aakanksha, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, et al. 2022. “PaLM: Scaling Language Modeling with Pathways.” arXiv. https://doi.org/10.48550/ARXIV.2204.02311.
Chowdhury, Ahmadul Karim, Md. Saidur Rahman Sujon, Md. Shirajus Salekin Shafi, Tasin Ahmmad, Sifat Ahmed, Khan Md Hasib, and Faisal Muhammad Shah. 2024. “Harnessing Large Language Models over Transformer Models for Detecting Bengali Depressive Social Media Text: A Comprehensive Study.” arXiv. https://doi.org/10.48550/ARXIV.2401.07310.
Christakopoulou, Evangelia, and George Karypis. 2016. “Local Item-Item Models for Top-n Recommendation.” Proceedings of the 10th ACM Conference on Recommender Systems, September. https://doi.org/10.1145/2959100.2959185.
Christensen, Dennis. 2024. “Inference for Bayesian Nonparametric Models with Binary Response Data via Permutation Counting.” Bayesian Analysis 19 (March). https://doi.org/10.1214/22-ba1353.
Christopoulou, Fenia, Makoto Miwa, and Sophia Ananiadou. 2019. “Connecting the Dots: Document-Level Neural Relation Extraction with Edge-Oriented Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1909.00228.
Chu, Xiangxiang, Bo Zhang, Qingyuan Li, Ruijun Xu, and Xudong Li. 2019. “SCARLET-NAS: Bridging the Gap Between Stability and Scalability in Weight-Sharing Neural Architecture Search.” arXiv. https://doi.org/10.48550/ARXIV.1908.06022.
Chuang, Ching-Yao, Varun Jampani, Yuanzhen Li, Antonio Torralba, and Stefanie Jegelka. 2023. “Debiasing Vision-Language Models via Biased Prompts.” arXiv. https://doi.org/10.48550/ARXIV.2302.00070.
Chuang, Ching-Yao, and Stefanie Jegelka. 2022. “Tree Mover’s Distance: Bridging Graph Metrics and Stability of Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2210.01906.
Chuang, Sheuwen, Kuo-Song Chang, David D. Woods, Hsiao-Chun Chen, Morgan E. Reynolds, and Ding-Kuo Chien. 2019. “Beyond Surge: Coping with Mass Burn Casualty in the Closest Hospital to the Formosa Fun Coast Dust Explosion.” Burns 45 (June). https://doi.org/10.1016/j.burns.2018.12.003.
Chuang, Yu-Neng, Ruixiang Tang, Xiaoqian Jiang, and Xia Hu. 2023. “SPeC: A Soft Prompt-Based Calibration on Performance Variability of Large Language Model in Clinical Notes Summarization.” arXiv. https://doi.org/10.48550/ARXIV.2303.13035.
Chung, Hyung Won, Le Hou, Shayne Longpre, Barret Zoph, Yi Tay, William Fedus, Yunxuan Li, et al. 2022. “Scaling Instruction-Finetuned Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2210.11416.
Chung, Joon Son, Arsha Nagrani, Ernesto Coto, Weidi Xie, Mitchell McLaren, Douglas A Reynolds, and Andrew Zisserman. 2019. “VoxSRC 2019: The First VoxCeleb Speaker Recognition Challenge.” arXiv. https://doi.org/10.48550/ARXIV.1912.02522.
Chung, Junyoung, Kyunghyun Cho, and Yoshua Bengio. 2016. “A Character-Level Decoder Without Explicit Segmentation for Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1603.06147.
Chung, Junyoung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. 2014. “Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling.” arXiv. https://doi.org/10.48550/ARXIV.1412.3555.
Çiflikli, Gökhan, Nils W Metternich, Sigrid Weber, and Kit Rickard. 2019. “Taking Time Seriously When Evaluating Predictions in Binary-Time-Series-Cross-Section-Data,” March. https://doi.org/10.31235/osf.io/tvshu.
Cireşan, Dan C., Ueli Meier, Jonathan Masci, Luca M. Gambardella, and Jürgen Schmidhuber. 2011. “High-Performance Neural Networks for Visual Object Classification.” arXiv. https://doi.org/10.48550/ARXIV.1102.0183.
Cireşan, Dan, Ueli Meier, and Juergen Schmidhuber. 2012. “Multi-Column Deep Neural Networks for Image Classification.” arXiv. https://doi.org/10.48550/ARXIV.1202.2745.
Cirstea, Razvan-Gabriel, Tung Kieu, Chenjuan Guo, Bin Yang, and Sinno Jialin Pan. 2021. “EnhanceNet: Plugin Neural Networks for Enhancing Correlated Time Series Forecasting.” 2021 IEEE 37th International Conference on Data Engineering (ICDE), April. https://doi.org/10.1109/icde51399.2021.00153.
Cisse, Moustapha, Piotr Bojanowski, Edouard Grave, Yann Dauphin, and Nicolas Usunier. 2017. “Parseval Networks: Improving Robustness to Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1704.08847.
Clark, Alexander. 2003. “Combining Distributional and Morphological Information for Part of Speech Induction.” Proceedings of the Tenth Conference on European Chapter of the Association for Computational Linguistics - EACL ’03. https://doi.org/10.3115/1067807.1067817.
Clark, Andy, and Josefa Toribio. 1994. “Doing Without Representing?” Synthese 101 (December). https://doi.org/10.1007/bf01063896.
Clark, K. L. 1980. “Algorithm Classification Through Synthesis.” The Computer Journal 23 (January). https://doi.org/10.1093/comjnl/23.1.61.
Clark, Peter, Isaac Cowhey, Oren Etzioni, Tushar Khot, Ashish Sabharwal, Carissa Schoenick, and Oyvind Tafjord. 2018. “Think You Have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge.” arXiv. https://doi.org/10.48550/ARXIV.1803.05457.
Clark, Peter, and Tim Niblett. 1989. “The CN2 Induction Algorithm.” Machine Learning 3 (March). https://doi.org/10.1007/bf00116835.
“Classic Works of the Dempster-Shafer Theory of Belief Functions.” 2008. Studies in Fuzziness and Soft Computing. https://doi.org/10.1007/978-3-540-44792-4.
Clevert, Djork-Arné, Thomas Unterthiner, and Sepp Hochreiter. 2015. “Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs).” arXiv. https://doi.org/10.48550/ARXIV.1511.07289.
Clouâtre, Louis, and Marc Demers. 2019. “FIGR: Few-Shot Image Generation with Reptile.” arXiv. https://doi.org/10.48550/ARXIV.1901.02199.
Clune, Jeff. 2019. “AI-GAs: AI-Generating Algorithms, an Alternate Paradigm for Producing General Artificial Intelligence.” arXiv. https://doi.org/10.48550/ARXIV.1905.10985.
Coda-Forno, Julian, Kristin Witte, Akshay K. Jagadish, Marcel Binz, Zeynep Akata, and Eric Schulz. 2023. “Inducing Anxiety in Large Language Models Increases Exploration and Bias.” arXiv. https://doi.org/10.48550/ARXIV.2304.11111.
Cohan, Arman, Sergey Feldman, Iz Beltagy, Doug Downey, and Daniel S. Weld. 2020. “SPECTER: Document-Level Representation Learning Using Citation-Informed Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2004.07180.
Cohen, Roi, Mor Geva, Jonathan Berant, and Amir Globerson. 2023. “Crawling the Internal Knowledge-Base of Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.12810.
Cohen, Taco, Mario Geiger, Jonas Köhler, and Max Welling. 2017. “Convolutional Networks for Spherical Signals.” arXiv. https://doi.org/10.48550/ARXIV.1709.04893.
Cohen, William W. 2016. “TensorLog: A Differentiable Deductive Database.” arXiv. https://doi.org/10.48550/ARXIV.1605.06523.
Collins, Liam, Aryan Mokhtari, and Sanjay Shakkottai. 2020. “How Does the Task Landscape Affect MAML Performance?” arXiv. https://doi.org/10.48550/ARXIV.2010.14672.
Collobert, Ronan, Christian Puhrsch, and Gabriel Synnaeve. 2016. “Wav2Letter: An End-to-End ConvNet-Based Speech Recognition System.” arXiv. https://doi.org/10.48550/ARXIV.1609.03193.
Collobert, Ronan, and Jason Weston. 2008. “A Unified Architecture for Natural Language Processing.” Proceedings of the 25th International Conference on Machine Learning - ICML ’08. https://doi.org/10.1145/1390156.1390177.
Collobert, Ronan, Jason Weston, Leon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. “Natural Language Processing (Almost) from Scratch.” arXiv. https://doi.org/10.48550/ARXIV.1103.0398.
Colombo, Diego, and Marloes H. Maathuis. 2012. “Order-Independent Constraint-Based Causal Structure Learning.” arXiv. https://doi.org/10.48550/ARXIV.1211.3295.
Comelli, Thibaud, Frédéric Pinel, and Pascal Bouvry. 2021. “Comparing Elementary Cellular Automata Classifications with a Convolutional Neural Network.” Proceedings of the 13th International Conference on Agents and Artificial Intelligence. https://doi.org/10.5220/0010160004670474.
“Complex Automated Negotiations: Theories, Models, and Software Competitions.” 2013. Studies in Computational Intelligence. https://doi.org/10.1007/978-3-642-30737-9.
“Complex Networks &Amp; Their Applications v.” 2017. Studies in Computational Intelligence. https://doi.org/10.1007/978-3-319-50901-3.
“Computer Vision – ACCV 2018.” 2019. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-20893-6.
Congedo, Marco, Alexandre Barachant, and Anton Andreev. 2013. “A New Generation of Brain-Computer Interface Based on Riemannian Geometry.” arXiv. https://doi.org/10.48550/ARXIV.1310.8115.
Conneau, Alexis, German Kruszewski, Guillaume Lample, Loïc Barrault, and Marco Baroni. 2018. “What You Can Cram into a Single Vector: Probing Sentence Embeddings for Linguistic Properties.” arXiv. https://doi.org/10.48550/ARXIV.1805.01070.
Conneau, Alexis, Min Ma, Simran Khanuja, Yu Zhang, Vera Axelrod, Siddharth Dalmia, Jason Riesa, Clara Rivera, and Ankur Bapna. 2022. “FLEURS: Few-Shot Learning Evaluation of Universal Representations of Speech.” arXiv. https://doi.org/10.48550/ARXIV.2205.12446.
Conneau, Alexis, Ruty Rinott, Guillaume Lample, Adina Williams, Samuel Bowman, Holger Schwenk, and Veselin Stoyanov. 2018. “XNLI: Evaluating Cross-Lingual Sentence Representations.” Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/d18-1269.
Conneau, Alexis, Holger Schwenk, Loïc Barrault, and Yann Lecun. 2016. “Very Deep Convolutional Networks for Text Classification.” arXiv. https://doi.org/10.48550/ARXIV.1606.01781.
Cook, Samantha R, Andrew Gelman, and Donald B Rubin. 2006. “Validation of Software for Bayesian Models Using Posterior Quantiles.” Journal of Computational and Graphical Statistics 15 (September). https://doi.org/10.1198/106186006x136976.
Coquelin, Pierre-Arnuad, and Remi Munos. 2014. “Bandit Algorithms for Tree Search.” arXiv. https://doi.org/10.48550/ARXIV.1408.2028.
Coquet, François, Ying Hu, Jean Mémin, and Shige Peng. 2002. “Filtration-Consistent Nonlinear Expectations and Related g-Expectations.” Probability Theory and Related Fields 123 (May). https://doi.org/10.1007/s004400100172.
Corneliou, Panayiotis, Panagiota Nikolaou, Maria K. Michael, and Theocharis Theocharides. 2021. “Fine-Grained Vulnerability Analysis of Resource Constrained Neural Inference Accelerators.” 2021 IEEE International Symposium on Defect and Fault Tolerance in VLSI and Nanotechnology Systems (DFT), October. https://doi.org/10.1109/dft52944.2021.9568281.
Cortes, Corinna, Mehryar Mohri, Michael Riley, and Afshin Rostamizadeh. 2008. “Sample Selection Bias Correction Theory.” arXiv. https://doi.org/10.48550/ARXIV.0805.2775.
Cortes, Corinna, Mehryar Mohri, and Afshin Rostamizadeh. 2012. “Algorithms for Learning Kernels Based on Centered Alignment.” arXiv. https://doi.org/10.48550/ARXIV.1203.0550.
Costabal, Francisco Sahli, Simone Pezzuto, and Paris Perdikaris. 2022. \(Δ\)-PINNs: Physics-Informed Neural Networks on Complex Geometries.” arXiv. https://doi.org/10.48550/ARXIV.2209.03984.
Cotter, Andrew, Ohad Shamir, Nathan Srebro, and Karthik Sridharan. 2011. “Better Mini-Batch Algorithms via Accelerated Gradient Methods.” arXiv. https://doi.org/10.48550/ARXIV.1106.4574.
Cotterell, Ryan, and Jason Eisner. 2017. “Probabilistic Typology: Deep Generative Models of Vowel Inventories.” arXiv. https://doi.org/10.48550/ARXIV.1705.01684.
Couairon, Guillaume, Jakob Verbeek, Holger Schwenk, and Matthieu Cord. 2022. “DiffEdit: Diffusion-Based Semantic Image Editing with Mask Guidance.” arXiv. https://doi.org/10.48550/ARXIV.2210.11427.
Coucke, Alice, Alaa Saade, Adrien Ball, Théodore Bluche, Alexandre Caulier, David Leroy, Clément Doumouro, et al. 2018. “Snips Voice Platform: An Embedded Spoken Language Understanding System for Private-by-Design Voice Interfaces.” arXiv. https://doi.org/10.48550/ARXIV.1805.10190.
Coulombe, Claude. 2018. “Text Data Augmentation Made Simple by Leveraging NLP Cloud APIs.” arXiv. https://doi.org/10.48550/ARXIV.1812.04718.
Courbariaux, Matthieu, Yoshua Bengio, and Jean-Pierre David. 2014. “Training Deep Neural Networks with Low Precision Multiplications.” arXiv. https://doi.org/10.48550/ARXIV.1412.7024.
———. 2015. “BinaryConnect: Training Deep Neural Networks with Binary Weights During Propagations.” arXiv. https://doi.org/10.48550/ARXIV.1511.00363.
Courbariaux, Matthieu, Itay Hubara, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. 2016. “Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1.” arXiv. https://doi.org/10.48550/ARXIV.1602.02830.
Courty, Nicolas, Rémi Flamary, Amaury Habrard, and Alain Rakotomamonjy. 2017. “Joint Distribution Optimal Transportation for Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1705.08848.
Coutin, Laure, and Zhongmin Qian. 2002. “Stochastic Analysis, Rough Path Analysis and Fractional Brownian Motions.” Probability Theory and Related Fields 122 (January). https://doi.org/10.1007/s004400100158.
Covington, Paul, Jay Adams, and Emre Sargin. 2016. “Deep Neural Networks for YouTube Recommendations.” Proceedings of the 10th ACM Conference on Recommender Systems, September. https://doi.org/10.1145/2959100.2959190.
Cox, I. J., M. L. Miller, T. P. Minka, T. V. Papathomas, and P. N. Yianilos. 2000. “The Bayesian Image Retrieval System, PicHunter: Theory, Implementation, and Psychophysical Experiments.” IEEE Transactions on Image Processing 9. https://doi.org/10.1109/83.817596.
Cox, Ingemar J. 1993. “A Review of Statistical Data Association Techniques for Motion Correspondence.” International Journal of Computer Vision 10 (February). https://doi.org/10.1007/bf01440847.
Craiu, Radu V., Ruobin Gong, and Xiao-Li Meng. 2023. “Six Statistical Senses.” Annual Review of Statistics and Its Application 10 (March). https://doi.org/10.1146/annurev-statistics-040220-015348.
Crauel, Hans, and Franco Flandoli. 1994. “Attractors for Random Dynamical Systems.” Probability Theory and Related Fields 100 (September). https://doi.org/10.1007/bf01193705.
Cremaschi, Andrea, Raffaele Argiento, Maria De Iorio, Cai Shirong, Yap Seng Chong, Michael Meaney, and Michelle Kee. 2023. “Seemingly Unrelated Multi-State Processes: A Bayesian Semiparametric Approach.” Bayesian Analysis 18 (September). https://doi.org/10.1214/22-ba1326.
Cremonesi, Paolo, Yehuda Koren, and Roberto Turrin. 2010. “Performance of Recommender Algorithms on Top-n Recommendation Tasks.” Proceedings of the Fourth ACM Conference on Recommender Systems, September. https://doi.org/10.1145/1864708.1864721.
Creswell, Antonia, Kai Arulkumaran, and Anil Anthony Bharath. 2016. “Improving Sampling from Generative Autoencoders with Markov Chains.” arXiv. https://doi.org/10.48550/ARXIV.1610.09296.
Crowley, Elliot J., Gavin Gray, and Amos Storkey. 2017. “Moonshine: Distilling with Cheap Convolutions.” arXiv. https://doi.org/10.48550/ARXIV.1711.02613.
Csurka, Gabriela. 2017. “Domain Adaptation for Visual Applications: A Comprehensive Survey.” arXiv. https://doi.org/10.48550/ARXIV.1702.05374.
Cubuk, Ekin D., Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V. Le. 2019. “AutoAugment: Learning Augmentation Strategies from Data.” 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr.2019.00020.
Cui, Can, Yunsheng Ma, Xu Cao, Wenqian Ye, and Ziran Wang. 2023. “Receive, Reason, and React: Drive as You Say with Large Language Models in Autonomous Vehicles.” arXiv. https://doi.org/10.48550/ARXIV.2310.08034.
Cui, Chunfeng, Kaiqi Zhang, Talgat Daulbaev, Julia Gusak, Ivan Oseledets, and Zheng Zhang. 2019. “Active Subspace of Neural Networks: Structural Analysis and Universal Attacks.” arXiv. https://doi.org/10.48550/ARXIV.1910.13025.
Cui, Jian, Kwanwoo Kim, Seung Ho Na, and Seungwon Shin. 2021. “Meta-Path-Based Fake News Detection Leveraging Multi-Level Social Context Information.” arXiv. https://doi.org/10.48550/ARXIV.2109.08022.
Cui, Jiaxu, Bo Yang, and Xia Hu. 2019. “Deep Bayesian Optimization on Attributed Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1905.13403.
Cui, Limeng, Xianfeng Tang, Sumeet Katariya, Nikhil Rao, Pallav Agrawal, Karthik Subbian, and Dongwon Lee. 2022. “ALLIE: Active Learning on Large-Scale Imbalanced Graphs.” Proceedings of the ACM Web Conference 2022, April. https://doi.org/10.1145/3485447.3512229.
Cui, Peng, Xiao Wang, Jian Pei, and Wenwu Zhu. 2017. “A Survey on Network Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1711.08752.
Cui, Yue, Kai Zheng, Dingshan Cui, Jiandong Xie, Liwei Deng, Feiteng Huang, and Xiaofang Zhou. 2021. “METRO.” Proceedings of the VLDB Endowment 15 (October). https://doi.org/10.14778/3489496.3489503.
Cui, Zeyu, Jianxin Ma, Chang Zhou, Jingren Zhou, and Hongxia Yang. 2022. “M6-Rec: Generative Pretrained Language Models Are Open-Ended Recommender Systems.” arXiv. https://doi.org/10.48550/ARXIV.2205.08084.
Čuljak, Marko, Andreas Spitz, Robert West, and Akhil Arora. 2022. “Strong Heuristics for Named Entity Linking.” arXiv. https://doi.org/10.48550/ARXIV.2207.02824.
Cummins, Chris, Pavlos Petoumenos, Alastair Murray, and Hugh Leather. 2018. “Compiler Fuzzing Through Deep Learning.” Proceedings of the 27th ACM SIGSOFT International Symposium on Software Testing and Analysis, July. https://doi.org/10.1145/3213846.3213848.
Curless, Brian, and Marc Levoy. 1996. “A Volumetric Method for Building Complex Models from Range Images.” Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, August. https://doi.org/10.1145/237170.237269.
Cussens, James. 2012. “Bayesian Network Learning with Cutting Planes.” arXiv. https://doi.org/10.48550/ARXIV.1202.3713.
Cusumano-Towner, Marco F., Feras A. Saad, Alexander K. Lew, and Vikash K. Mansinghka. 2019. “Gen: A General-Purpose Probabilistic Programming System with Programmable Inference.” Proceedings of the 40th ACM SIGPLAN Conference on Programming Language Design and Implementation, June. https://doi.org/10.1145/3314221.3314642.
Cuturi, Marco, and Mathieu Blondel. 2017. “Soft-DTW: A Differentiable Loss Function for Time-Series.” arXiv. https://doi.org/10.48550/ARXIV.1703.01541.
D’Amour, Alexander, Peng Ding, Avi Feller, Lihua Lei, and Jasjeet Sekhon. 2017. “Overlap in Observational Studies with High-Dimensional Covariates.” arXiv. https://doi.org/10.48550/ARXIV.1711.02582.
D’Amour, Alexander, Katherine Heller, Dan Moldovan, Ben Adlam, Babak Alipanahi, Alex Beutel, Christina Chen, et al. 2020. “Underspecification Presents Challenges for Credibility in Modern Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.2011.03395.
Dabrowski, Joel Janek, YiFan Zhang, and Ashfaqur Rahman. 2020. “ForecastNet: A Time-Variant Deep Feed-Forward Neural Network Architecture for Multi-Step-Ahead Time-Series Forecasting.” Neural Information Processing. https://doi.org/10.1007/978-3-030-63836-8_48.
Daelemans, Walter, Antal van den Bosch, and Jakub Zavrel. 1998. “Forgetting Exceptions Is Harmful in Language Learning.” arXiv. https://doi.org/10.48550/ARXIV.CS/9812021.
Daghero, Francesco, Daniele Jahier Pagliari, and Massimo Poncino. 2022. “Two-Stage Human Activity Recognition on Microcontrollers with Decision Trees and CNNs.” arXiv. https://doi.org/10.48550/ARXIV.2206.07652.
Dahl, Christian M., Torben Johansen, Emil N. Sørensen, and Simon Wittrock. 2021. “HANA: A HAndwritten NAme Database for Offline Handwritten Text Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2101.10862.
Dahl, G. E., Dong Yu, Li Deng, and A. Acero. 2012. “Context-Dependent Pre-Trained Deep Neural Networks for Large-Vocabulary Speech Recognition.” IEEE Transactions on Audio, Speech, and Language Processing 20 (January). https://doi.org/10.1109/tasl.2011.2134090.
Dai, Andrew M., and Quoc V. Le. 2015. “Semi-Supervised Sequence Learning.” arXiv. https://doi.org/10.48550/ARXIV.1511.01432.
Dai, Bo, Sanja Fidler, Raquel Urtasun, and Dahua Lin. 2017. “Towards Diverse and Natural Image Descriptions via a Conditional GAN.” arXiv. https://doi.org/10.48550/ARXIV.1703.06029.
Dai, Damai, Li Dong, Yaru Hao, Zhifang Sui, Baobao Chang, and Furu Wei. 2021. “Knowledge Neurons in Pretrained Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2104.08696.
Dai, Hanjun, Bo Dai, and Le Song. 2016. “Discriminative Embeddings of Latent Variable Models for Structured Data.” arXiv. https://doi.org/10.48550/ARXIV.1603.05629.
Dai, Hanjun, Hui Li, Tian Tian, Xin Huang, Lin Wang, Jun Zhu, and Le Song. 2018. “Adversarial Attack on Graph Structured Data.” arXiv. https://doi.org/10.48550/ARXIV.1806.02371.
Dai, Hanjun, Yichen Wang, Rakshit Trivedi, and Le Song. 2016. “Deep Coevolutionary Network: Embedding User and Item Features for Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.1609.03675.
Dai, Jifeng, Kaiming He, Yi Li, Shaoqing Ren, and Jian Sun. 2016. “Instance-Sensitive Fully Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1603.08678.
Dai, Jifeng, Haozhi Qi, Yuwen Xiong, Yi Li, Guodong Zhang, Han Hu, and Yichen Wei. 2017. “Deformable Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1703.06211.
Dai, Quanyu, Qiang Li, Jian Tang, and Dan Wang. 2017. “Adversarial Network Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1711.07838.
Dai, Quanyu, Xiao Shen, Liang Zhang, Qiang Li, and Dan Wang. 2019. “Adversarial Training Methods for Network Embedding.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313445.
Dai, Rui, Shenkun Xu, Qian Gu, Chenguang Ji, and Kaikui Liu. 2020. “Hybrid Spatio-Temporal Graph Convolutional Network.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403358.
Dai, Wei, and Olgica Milenkovic. 2008. “Subspace Pursuit for Compressive Sensing Signal Reconstruction.” arXiv. https://doi.org/10.48550/ARXIV.0803.0811.
Dai, Ya, Liang Guo, Steve Liu, and Hongxian Zhang. 2022. “Are Socially Responsible <Scp>exchange‐traded</Scp> Funds Paying Off in Performance?” International Review of Finance 23 (September). https://doi.org/10.1111/irfi.12389.
Dai, Yaoyao, Benjamin Radford, and Andrew Halterman. 2022. “Political Event Coding as Text-to-Text Sequence Generation.” Proceedings of the 5th Workshop on Challenges and Applications of Automated Extraction of Socio-Political Events from Text (CASE). https://doi.org/10.18653/v1/2022.case-1.16.
Dai, Zihang, Lei Li, and Wei Xu. 2016. “CFO: Conditional Focused Neural Question Answering with Large-Scale Knowledge Bases.” arXiv. https://doi.org/10.48550/ARXIV.1606.01994.
Daian, Philip, Steven Goldfeder, Tyler Kell, Yunqi Li, Xueyuan Zhao, Iddo Bentov, Lorenz Breidenbach, and Ari Juels. 2019. “Flash Boys 2.0: Frontrunning, Transaction Reordering, and Consensus Instability in Decentralized Exchanges.” arXiv. https://doi.org/10.48550/ARXIV.1904.05234.
Dang, Hai, Lukas Mecke, Florian Lehmann, Sven Goller, and Daniel Buschek. 2022. “How to Prompt? Opportunities and Challenges of Zero- and Few-Shot Learning for Human-AI Interaction in Creative Applications of Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.2209.01390.
Das, Abhinandan S., Mayur Datar, Ashutosh Garg, and Shyam Rajaram. 2007. “Google News Personalization.” Proceedings of the 16th International Conference on World Wide Web, May. https://doi.org/10.1145/1242572.1242610.
Das, Abhishek, Satwik Kottur, José M. F. Moura, Stefan Lee, and Dhruv Batra. 2017. “Learning Cooperative Visual Dialog Agents with Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1703.06585.
Das, Rajarshi, Shehzaad Dhuliawala, Manzil Zaheer, Luke Vilnis, Ishan Durugkar, Akshay Krishnamurthy, Alex Smola, and Andrew McCallum. 2017. “Go for a Walk and Arrive at the Answer: Reasoning over Paths in Knowledge Bases Using Reinforcement Learning,” November. http://arxiv.org/abs/1711.05851v2.
Dash, Ayushman, John Cristian Borges Gamboa, Sheraz Ahmed, Marcus Liwicki, and Muhammad Zeshan Afzal. 2017. “TAC-GAN - Text Conditioned Auxiliary Classifier Generative Adversarial Network.” arXiv. https://doi.org/10.48550/ARXIV.1703.06412.
“Data Mining and Knowledge Discovery Handbook.” 2005. https://doi.org/10.1007/b107408.
Datta, Gourav, Souvik Kundu, Zihan Yin, Ravi Teja Lakkireddy, Joe Mathai, Ajey P. Jacob, Peter A. Beerel, and Akhilesh R. Jaiswal. 2022. “A Processing-in-Pixel-in-Memory Paradigm for Resource-Constrained TinyML Applications.” Scientific Reports 12 (August). https://doi.org/10.1038/s41598-022-17934-1.
Daume, Hal, Nikos Karampatziakis, John Langford, and Paul Mineiro. 2016. “Logarithmic Time One-Against-Some.” arXiv. https://doi.org/10.48550/ARXIV.1606.04988.
Daumé, Hal, John Langford, and Daniel Marcu. 2009. “Search-Based Structured Prediction.” arXiv. https://doi.org/10.48550/ARXIV.0907.0786.
Dauphin, Yann N., Harm de Vries, and Yoshua Bengio. 2015. “Equilibrated Adaptive Learning Rates for Non-Convex Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1502.04390.
Daus, Esther S., Mariya Ptashnyk, and Claudia Raithel. 2020. “Derivation of a Fractional Cross-Diffusion System as the Limit of a Stochastic Many-Particle System Driven by Lévy Noise.” arXiv. https://doi.org/10.48550/ARXIV.2006.00277.
David, Eli, Nathan S. Netanyahu, and Lior Wolf. 2017. “DeepChess: End-to-End Deep Neural Network for Automatic Learning in Chess.” arXiv. https://doi.org/10.48550/ARXIV.1711.09667.
Daw, Arka, M. Maruf, and Anuj Karpatne. 2021. “PID-GAN: A GAN Framework Based on a Physics-Informed Discriminator for Uncertainty Quantification with Physics.” Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3447548.3467449.
Dazzi, Martino, Abu Sebastian, Thomas Parnell, Pier Andrea Francese, Luca Benini, and Evangelos Eleftheriou. 2021. “Efficient Pipelined Execution of CNNs Based on in-Memory Computing and Graph Homomorphism Verification.” IEEE Transactions on Computers 70 (June). https://doi.org/10.1109/tc.2021.3073255.
Dean, Sarah, Horia Mania, Nikolai Matni, Benjamin Recht, and Stephen Tu. 2017. “On the Sample Complexity of the Linear Quadratic Regulator.” arXiv. https://doi.org/10.48550/ARXIV.1710.01688.
Debnath, Asim Kumar, Rosa L. Lopez de Compadre, Gargi Debnath, Alan J. Shusterman, and Corwin Hansch. 1991. “Structure-Activity Relationship of Mutagenic Aromatic and Heteroaromatic Nitro Compounds. Correlation with Molecular Orbital Energies and Hydrophobicity.” Journal of Medicinal Chemistry 34 (February). https://doi.org/10.1021/jm00106a046.
DeChant, Chad, Tyr Wiesner-Hanks, Siyuan Chen, Ethan L. Stewart, Jason Yosinski, Michael A. Gore, Rebecca J. Nelson, and Hod Lipson. 2017. “Automated Identification of Northern Leaf Blight-Infected Maize Plants from Field Imagery Using Deep Learning.” Phytopathology® 107 (November). https://doi.org/10.1094/phyto-11-16-0417-r.
Defazio, Aaron, and Konstantin Mishchenko. 2023. “Learning-Rate-Free Learning by d-Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.2301.07733.
Defferrard, Michaël, Xavier Bresson, and Pierre Vandergheynst. 2016. “Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering.” arXiv. https://doi.org/10.48550/ARXIV.1606.09375.
Dehghan, Afshin, Enrique G. Ortiz, Guang Shu, and Syed Zain Masood. 2017. “DAGER: Deep Age, Gender and Emotion Recognition Using Convolutional Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.1702.04280.
Dehghani, Mostafa, Aliaksei Severyn, Sascha Rothe, and Jaap Kamps. 2017. “Learning to Learn from Weak Supervision by Full Supervision.” arXiv. https://doi.org/10.48550/ARXIV.1711.11383.
Del-Agua, Miguel Angel, Adria Gimenez, Albert Sanchis, Jorge Civera, and Alfons Juan. 2018. “Speaker-Adapted Confidence Measures for ASR Using Deep Bidirectional Recurrent Neural Networks.” IEEE/ACM Transactions on Audio, Speech, and Language Processing 26 (July). https://doi.org/10.1109/taslp.2018.2819900.
Dell’Eva, Anthony, Marco Orsingher, and Massimo Bertozzi. 2022. “Arbitrary Point Cloud Upsampling with Spherical Mixture of Gaussians.” arXiv. https://doi.org/10.48550/ARXIV.2208.05274.
Delmas, Alberto, Sayeh Sharify, Patrick Judd, and Andreas Moshovos. 2017. “Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability.” arXiv. https://doi.org/10.48550/ARXIV.1707.09068.
Demrozi, Florenc, Cristian Turetta, Fabio Chiarani, Philipp H. Kindt, and Graziano Pravadelli. 2021. “Estimating Indoor Occupancy Through Low-Cost BLE Devices.” IEEE Sensors Journal 21 (August). https://doi.org/10.1109/jsen.2021.3080632.
Deng, Jia, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. “ImageNet: A Large-Scale Hierarchical Image Database.” 2009 IEEE Conference on Computer Vision and Pattern Recognition, June. https://doi.org/10.1109/cvpr.2009.5206848.
Deng, Li. 2014. “Deep Learning: Methods and Applications.” Foundations and Trends® in Signal Processing 7. https://doi.org/10.1561/2000000039.
Deng, L., Michael L. Seltzer, Dong Yu, Alex Acero, Abdel-rahman Mohamed, and G. Hinton. 2010. “Binary Coding of Speech Spectrograms Using a Deep Auto-Encoder.” Interspeech 2010, September. https://doi.org/10.21437/interspeech.2010-487.
Deng, Songgaojun, and Yue Ning. 2021. “A Survey on Societal Event Forecasting with Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.2112.06345.
Deng, Yichuan, Zhihang Li, and Zhao Song. 2023. “Attention Scheme Inspired Softmax Regression.” arXiv. https://doi.org/10.48550/ARXIV.2304.10411.
Deng, Yongshi, and Thomas Lumley. 2021. “Multiple Imputation Through XGBoost.” arXiv. https://doi.org/10.48550/ARXIV.2106.01574.
Denison, D. G. T., and C. C. Holmes. 2001. “Bayesian Partitioning for Estimating Disease Risk.” Biometrics 57 (March). https://doi.org/10.1111/j.0006-341x.2001.00143.x.
Denison, D. G. T., B. K. Mallick, and A. F. M. Smith. 1998. “Automatic Bayesian Curve Fitting.” Journal of the Royal Statistical Society Series B: Statistical Methodology 60 (July). https://doi.org/10.1111/1467-9868.00128.
Denton, Emily, Soumith Chintala, Arthur Szlam, and Rob Fergus. 2015. “Deep Generative Image Models Using a Laplacian Pyramid of Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1506.05751.
Denton, Emily, Sam Gross, and Rob Fergus. 2016. “Semi-Supervised Learning with Context-Conditional Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.06430.
Denton, Emily, Wojciech Zaremba, Joan Bruna, Yann LeCun, and Rob Fergus. 2014. “Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation.” arXiv. https://doi.org/10.48550/ARXIV.1404.0736.
Desai, Aditya, Sumit Gulwani, Vineet Hingorani, Nidhi Jain, Amey Karkare, Mark Marron, Sailesh R, and Subhajit Roy. 2016. “Program Synthesis Using Natural Language.” Proceedings of the 38th International Conference on Software Engineering, May. https://doi.org/10.1145/2884781.2884786.
DeSantis, Derek, Phillip J. Wolfram, Katrina Bennett, and Boian Alexandrov. 2020. “Coarse-Grain Cluster Analysis of Tensors with Application to Climate Biome Identification.” arXiv. https://doi.org/10.48550/ARXIV.2001.07827.
Desjardins, Guillaume, Aaron Courville, and Yoshua Bengio. 2012. “Disentangling Factors of Variation via Generative Entangling.” arXiv. https://doi.org/10.48550/ARXIV.1210.5474.
Dettmers, Tim. 2015. “8-Bit Approximations for Parallelism in Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1511.04561.
Dettmers, Tim, Mike Lewis, Younes Belkada, and Luke Zettlemoyer. 2022. “LLM.int8(): 8-Bit Matrix Multiplication for Transformers at Scale.” arXiv. https://doi.org/10.48550/ARXIV.2208.07339.
Dettmers, Tim, Pasquale Minervini, Pontus Stenetorp, and Sebastian Riedel. 2017. “Convolutional 2D Knowledge Graph Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1707.01476.
Dettmers, Tim, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer. 2023. “QLoRA: Efficient Finetuning of Quantized LLMs.” arXiv. https://doi.org/10.48550/ARXIV.2305.14314.
Devlin, Jacob, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. “BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding,” October. http://arxiv.org/abs/1810.04805v2.
———. 2019. Proceedings of the 2019 Conference of the North. https://doi.org/10.18653/v1/n19-1423.
DeVries, Terrance, and Graham W. Taylor. 2017. “Improved Regularization of Convolutional Neural Networks with Cutout.” arXiv. https://doi.org/10.48550/ARXIV.1708.04552.
Dhaou, Amin, Antoine Bertoncello, Sébastien Gourvénec, Josselin Garnier, and Erwan Le Pennec. 2021. “Causal and Interpretable Rules for Time Series Analysis.” Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3447548.3467161.
Dhawan, Sarthika, Siva Charan Reddy Gangireddy, Shiv Kumar, and Tanmoy Chakraborty. 2019. “Spotting Collective Behaviour of Online Frauds in Customer Reviews.” arXiv. https://doi.org/10.48550/ARXIV.1905.13649.
Dhingra, Bhuwan, Hanxiao Liu, Zhilin Yang, William W. Cohen, and Ruslan Salakhutdinov. 2016. “Gated-Attention Readers for Text Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.1606.01549.
Dhurandhar, Amit, Vijay Iyengar, Ronny Luss, and Karthikeyan Shanmugam. 2017. “TIP: Typifying the Interpretability of Procedures.” arXiv. https://doi.org/10.48550/ARXIV.1706.02952.
Diamond, Steven, and Stephen Boyd. 2016. “CVXPY: A Python-Embedded Modeling Language for Convex Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1603.00943.
Diamond, Steven, Vincent Sitzmann, Felix Heide, and Gordon Wetzstein. 2017. “Unrolled Optimization with Deep Priors.” arXiv. https://doi.org/10.48550/ARXIV.1705.08041.
Diao, Cameron, Kaixiong Zhou, Zirui Liu, Xiao Huang, and Xia Hu. 2022. “MolCPT: Molecule Continuous Prompt Tuning to Generalize Molecular Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.2212.10614.
Díaz, Agustín Ortíz, José del Campo-Ávila, Gonzalo Ramos-Jiménez, Isvani Frías Blanco, Yailé Caballero Mota, Antonio Mustelier Hechavarría, and Rafael Morales-Bueno. 2015. “Fast Adapting Ensemble: A New Algorithm for Mining Data Streams with Concept Drift.” The Scientific World Journal 2015. https://doi.org/10.1155/2015/235810.
Diba, Ali, Mohsen Fayyaz, Vivek Sharma, Amir Hossein Karami, Mohammad Mahdi Arzani, Rahman Yousefzadeh, and Luc Van Gool. 2017. “Temporal 3D ConvNets: New Architecture and Transfer Learning for Video Classification.” arXiv. https://doi.org/10.48550/ARXIV.1711.08200.
Diehl, Peter U., and Matthew Cook. 2015. “Unsupervised Learning of Digit Recognition Using Spike-Timing-Dependent Plasticity.” Frontiers in Computational Neuroscience 9 (August). https://doi.org/10.3389/fncom.2015.00099.
Dieleman, Sander, Jan Schlüter, Colin Raffel, Søren Kaae Sønderby, Daniel Nouri, Daniel Maturana, Martin Thoma, et al. 2015. “Lasagne: First Release.” August. https://doi.org/10.5281/ZENODO.27878.
Diemert, Eustache, Julien Meynet, Pierre Galland, and Damien Lefortier. 2017. “Attribution Modeling Increases Efficiency of Bidding in Display Advertising.” arXiv. https://doi.org/10.48550/ARXIV.1707.06409.
Dietterich, Thomas G. 1999. “Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition.” arXiv. https://doi.org/10.48550/ARXIV.CS/9905014.
Dietterich, Thomas G., Adam Ashenfelter, and Yaroslav Bulatov. 2004. “Training Conditional Random Fields via Gradient Tree Boosting.” Twenty-First International Conference on Machine Learning - ICML ’04. https://doi.org/10.1145/1015330.1015428.
Dieuleveut, Aymeric, Alain Durmus, and Francis Bach. 2017. “Bridging the Gap Between Constant Step Size Stochastic Gradient Descent and Markov Chains.” arXiv. https://doi.org/10.48550/ARXIV.1707.06386.
Dimakopoulou, Maria, Nikos Vlassis, and Tony Jebara. 2019. “Marginal Posterior Sampling for Slate Bandits.” Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, August. https://doi.org/10.24963/ijcai.2019/308.
Dimitrakakis, Christos, Blaine Nelson, and Zuhe Zhang, Aikaterini Mitrokotsa, and Benjamin Rubinstein. 2013. “Bayesian Differential Privacy Through Posterior Sampling.” arXiv. https://doi.org/10.48550/ARXIV.1306.1066.
Ding, Caiwen, Siyu Liao, Yanzhi Wang, Zhe Li, Ning Liu, Youwei Zhuo, Chao Wang, et al. 2017. “C <Scp>ir</Scp> CNN.” Proceedings of the 50th Annual IEEE/ACM International Symposium on Microarchitecture, October. https://doi.org/10.1145/3123939.3124552.
Ding, Caiwen, Shuo Wang, Ning Liu, Kaidi Xu, Yanzhi Wang, and Yun Liang. 2019. “REQ-YOLO: A Resource-Aware, Efficient Quantization Framework for Object Detection on FPGAs.” arXiv. https://doi.org/10.48550/ARXIV.1909.13396.
Ding, Daizong, Mi Zhang, Xudong Pan, Min Yang, and Xiangnan He. 2019. “Modeling Extreme Events in Time Series Prediction.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330896.
Ding, Hao, Changchang Sun, Hao Tang, Dawen Cai, and Yan Yan. 2022. “Few-Shot Medical Image Segmentation with Cycle-Resemblance Attention.” arXiv. https://doi.org/10.48550/ARXIV.2212.03967.
Ding, Kaize, Xuan Shan, and Huan Liu. 2021. “Towards Anomaly-Resistant Graph Neural Networks via Reinforcement Learning.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3482203.
Ding, Ning, Shengding Hu, Weilin Zhao, Yulin Chen, Zhiyuan Liu, Hai-Tao Zheng, and Maosong Sun. 2021. “OpenPrompt: An Open-Source Framework for Prompt-Learning.” arXiv. https://doi.org/10.48550/ARXIV.2111.01998.
Ding, Ning, Yujia Qin, Guang Yang, Fuchao Wei, Zonghan Yang, Yusheng Su, Shengding Hu, et al. 2022. “Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for Pre-Trained Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2203.06904.
Ding, Peng. 2023. “A First Course in Causal Inference.” arXiv. https://doi.org/10.48550/ARXIV.2305.18793.
Ding, Pengfei, Guanfeng Liu, Pengpeng Zhao, An Liu, Zhixu Li, and Kai Zheng. 2019. “Reinforcement Learning Based Monte Carlo Tree Search for Temporal Path Discovery.” 2019 IEEE International Conference on Data Mining (ICDM), November. https://doi.org/10.1109/icdm.2019.00024.
Ding, Qinxu, Yong Liu, Chunyan Miao, Fei Cheng, and Haihong Tang. 2020. “A Hybrid Bandit Framework for Diversified Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.2012.13245.
Ding, Xiaohan, Guiguang Ding, Xiangxin Zhou, Yuchen Guo, Jungong Han, and Ji Liu. 2019. “Global Sparse Momentum SGD for Pruning Very Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1909.12778.
Ding, Yan, Xiaohan Zhang, Chris Paxton, and Shiqi Zhang. 2023. “Task and Motion Planning with Large Language Models for Object Rearrangement.” arXiv. https://doi.org/10.48550/ARXIV.2303.06247.
Ding, Zhihao, Jieming Shi, Qing Li, and Jiannong Cao. 2023. “Effective Multi-Graph Neural Networks for Illicit Account Detection on Cryptocurrency Transaction Networks.” arXiv. https://doi.org/10.48550/ARXIV.2309.02460.
Dingliwal, Saket, Monica Sunkara, Sravan Bodapati, Srikanth Ronanki, Jeff Farris, and Katrin Kirchhoff. 2022. “Towards Personalization of CTC Speech Recognition Models with Contextual Adapters and Adaptive Boosting.” arXiv. https://doi.org/10.48550/ARXIV.2210.09510.
Dinh, Laurent, David Krueger, and Yoshua Bengio. 2014. “NICE: Non-Linear Independent Components Estimation.” arXiv. https://doi.org/10.48550/ARXIV.1410.8516.
Dinh, Laurent, Jascha Sohl-Dickstein, and Samy Bengio. 2016. “Density Estimation Using Real NVP.” arXiv. https://doi.org/10.48550/ARXIV.1605.08803.
Dinu, Georgiana, Angeliki Lazaridou, and Marco Baroni. 2014. “Improving Zero-Shot Learning by Mitigating the Hubness Problem.” arXiv. https://doi.org/10.48550/ARXIV.1412.6568.
Dis, Eva A. M. van, Johan Bollen, Willem Zuidema, Robert van Rooij, and Claudi L. Bockting. 2023. “ChatGPT: Five Priorities for Research.” Nature 614 (February). https://doi.org/10.1038/d41586-023-00288-7.
“Distill.” n.d. https://doi.org/10.23915/distill.
Do, Kien, Truyen Tran, and Svetha Venkatesh. 2018. “Graph Transformation Policy Network for Chemical Reaction Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1812.09441.
Do, Phong-Khac, Huy-Tien Nguyen, Chien-Xuan Tran, Minh-Tien Nguyen, and Minh-Le Nguyen. 2017. “Legal Question Answering Using Ranking SVM and Deep Convolutional Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.1703.05320.
Do, Xuan Long, Yiran Zhao, Hannah Brown, Yuxi Xie, James Xu Zhao, Nancy F. Chen, Kenji Kawaguchi, Michael Qizhe Xie, and Junxian He. 2023. “Prompt Optimization via Adversarial in-Context Learning.” arXiv. https://doi.org/10.48550/ARXIV.2312.02614.
“Document Analysis and Recognition – ICDAR 2021.” 2021. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-86331-9.
Dodds, Peter Sheridan, Kameron Decker Harris, Isabel M. Kloumann, Catherine A. Bliss, and Christopher M. Danforth. 2011. “Temporal Patterns of Happiness and Information in a Global Social Network: Hedonometrics and Twitter.” PLoS ONE 6 (December). https://doi.org/10.1371/journal.pone.0026752.
Dodge, Jesse, Gabriel Ilharco, Roy Schwartz, Ali Farhadi, Hannaneh Hajishirzi, and Noah Smith. 2020. “Fine-Tuning Pretrained Language Models: Weight Initializations, Data Orders, and Early Stopping.” arXiv. https://doi.org/10.48550/ARXIV.2002.06305.
Doersch, Carl. 2016. “Tutorial on Variational Autoencoders,” June. http://arxiv.org/abs/1606.05908v3.
Doersch, Carl, and Andrew Zisserman. 2017. “Multi-Task Self-Supervised Visual Learning.” arXiv. https://doi.org/10.48550/ARXIV.1708.07860.
Domingos, Pedro, and Michael Pazzani. 1997. Machine Learning 29. https://doi.org/10.1023/a:1007413511361.
Donahue, Chris, Bo Li, and Rohit Prabhavalkar. 2017. “Exploring Speech Enhancement with Generative Adversarial Networks for Robust Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1711.05747.
Donahue, Jeff, Yangqing Jia, Oriol Vinyals, Judy Hoffman, Ning Zhang, Eric Tzeng, and Trevor Darrell. 2013. “DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1310.1531.
Donahue, Jeff, Philipp Krähenbühl, and Trevor Darrell. 2016. “Adversarial Feature Learning.” arXiv. https://doi.org/10.48550/ARXIV.1605.09782.
Donahue, Jeff, and Karen Simonyan. 2019. “Large Scale Adversarial Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1907.02544.
Dong, Guozhu, and Jinyan Li. 1999. “Efficient Mining of Emerging Patterns.” Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/312129.312191.
Dong, Kun, Austin R. Benson, and David Bindel. 2019. “Network Density of States.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330891.
Dong, Li, and Mirella Lapata. 2018. “Coarse-to-Fine Decoding for Neural Semantic Parsing.” arXiv. https://doi.org/10.48550/ARXIV.1805.04793.
Dong, Li, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, and Hsiao-Wuen Hon. 2019. “Unified Language Model Pre-Training for Natural Language Understanding and Generation.” arXiv. https://doi.org/10.48550/ARXIV.1905.03197.
Dong, Xin Luna, Xiang He, Andrey Kan, Xian Li, Yan Liang, Jun Ma, Yifan Ethan Xu, et al. 2020. “AutoKnow.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403323.
Dong, Yihong, Kangcheng Luo, Xue Jiang, Zhi Jin, and Ge Li. 2023. “PACE: Improving Prompt with Actor-Critic Editing for Large Language Model.” arXiv. https://doi.org/10.48550/ARXIV.2308.10088.
Dong, Yue, Yikang Shen, Eric Crawford, Herke van Hoof, and Jackie Chi Kit Cheung. 2018. “BanditSum: Extractive Summarization as a Contextual Bandit.” arXiv. https://doi.org/10.48550/ARXIV.1809.09672.
Donnat, Claire, and Susan Holmes. 2018. “Tracking Network Dynamics: A Survey Using Graph Distances.” The Annals of Applied Statistics 12 (June). https://doi.org/10.1214/18-aoas1176.
Donnat, Claire, Marinka Zitnik, David Hallac, and Jure Leskovec. 2018. “Learning Structural Node Embeddings via Diffusion Wavelets.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220025.
Donoho, David L., and Iain M. Johnstone. 1995. “Adapting to Unknown Smoothness via Wavelet Shrinkage.” Journal of the American Statistical Association 90 (December). https://doi.org/10.1080/01621459.1995.10476626.
———. 1998. “Minimax Estimation via Wavelet Shrinkage.” The Annals of Statistics 26 (June). https://doi.org/10.1214/aos/1024691081.
Doran, Derek, Sarah Schulz, and Tarek R. Besold. 2017. “What Does Explainable AI Really Mean? A New Conceptualization of Perspectives.” arXiv. https://doi.org/10.48550/ARXIV.1710.00794.
Doshi-Velez, Finale, and Been Kim. 2017. “Towards a Rigorous Science of Interpretable Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1702.08608.
Doshi-Velez, Finale, Mason Kortz, Ryan Budish, Chris Bavitz, Sam Gershman, David O’Brien, Kate Scott, et al. 2017. “Accountability of AI Under the Law: The Role of Explanation.” arXiv. https://doi.org/10.48550/ARXIV.1711.01134.
Dosovitskiy, Alexey, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, et al. 2020. “An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale.” arXiv. https://doi.org/10.48550/ARXIV.2010.11929.
Dosovitskiy, Alexey, Philipp Fischer, Jost Tobias Springenberg, Martin Riedmiller, and Thomas Brox. 2016. “Discriminative Unsupervised Feature Learning with Exemplar Convolutional Neural Networks.” IEEE Transactions on Pattern Analysis and Machine Intelligence 38 (September). https://doi.org/10.1109/tpami.2015.2496141.
Dou, Wenwen, Li Yu, Xiaoyu Wang, Zhiqiang Ma, and William Ribarsky. 2013. “HierarchicalTopics: Visually Exploring Large Text Collections Using Topic Hierarchies.” IEEE Transactions on Visualization and Computer Graphics 19 (December). https://doi.org/10.1109/tvcg.2013.162.
Dou, Yingtong, Zhiwei Liu, Li Sun, Yutong Deng, Hao Peng, and Philip S. Yu. 2020. “Enhancing Graph Neural Network-Based Fraud Detectors Against Camouflaged Fraudsters.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3411903.
Dou, Yingtong, Kai Shu, Congying Xia, Philip S. Yu, and Lichao Sun. 2021. “User Preference-Aware Fake News Detection.” arXiv. https://doi.org/10.48550/ARXIV.2104.12259.
Douglas, B. L. 2011. “The Weisfeiler-Lehman Method and Graph Isomorphism Testing.” arXiv. https://doi.org/10.48550/ARXIV.1101.5211.
Douglass, Rex W, David A Meyer, Megha Ram, David Rideout, and Dongjin Song. 2015. “High Resolution Population Estimates from Telecommunications Data.” EPJ Data Science 4 (May). https://doi.org/10.1140/epjds/s13688-015-0040-6.
Doutre, Thibault, Wei Han, Min Ma, Zhiyun Lu, Chung-Cheng Chiu, Ruoming Pang, Arun Narayanan, Ananya Misra, Yu Zhang, and Liangliang Cao. 2020. “Improving Streaming Automatic Speech Recognition with Non-Streaming Model Distillation on Unsupervised Data.” arXiv. https://doi.org/10.48550/ARXIV.2010.12096.
Driess, Danny, Fei Xia, Mehdi S. M. Sajjadi, Corey Lynch, Aakanksha Chowdhery, Brian Ichter, Ayzaan Wahid, et al. 2023. “PaLM-e: An Embodied Multimodal Language Model.” arXiv. https://doi.org/10.48550/ARXIV.2303.03378.
Drovandi, Christopher, Richard G. Everitt, Andrew Golightly, and Dennis Prangle. 2022. “Ensemble MCMC: Accelerating Pseudo-Marginal MCMC for State Space Models Using the Ensemble Kalman Filter.” Bayesian Analysis 17 (March). https://doi.org/10.1214/20-ba1251.
Druck, Gregory, Gideon Mann, and Andrew McCallum. 2008. “Learning from Labeled Features Using Generalized Expectation Criteria.” Proceedings of the 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/1390334.1390436.
Du, Hongyang, Zonghang Li, Dusit Niyato, Jiawen Kang, Zehui Xiong, Xuemin, and Dong In Kim. 2023. “Enabling AI-Generated Content (AIGC) Services in Wireless Edge Networks.” arXiv. https://doi.org/10.48550/ARXIV.2301.03220.
Du, Jiawei, Hu Zhang, Joey Tianyi Zhou, Yi Yang, and Jiashi Feng. 2019. “Query-Efficient Meta Attack to Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1906.02398.
Du, Lun, Yun Wang, Guojie Song, Zhicong Lu, and Junshan Wang. 2018. “Dynamic Network Embedding : An Extended Approach for Skip-Gram Based Network Embedding.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/288.
Du, Mengnan, Ninghao Liu, and Xia Hu. 2018. “Techniques for Interpretable Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1808.00033.
Du, Min, Feifei Li, Guineng Zheng, and Vivek Srikumar. 2017. “DeepLog.” Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, October. https://doi.org/10.1145/3133956.3134015.
Du, Xinya, Junru Shao, and Claire Cardie. 2017. “Learning to Ask: Neural Question Generation for Reading Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.1705.00106.
Du, Yilun, Mengjiao Yang, Pete Florence, Fei Xia, Ayzaan Wahid, Brian Ichter, Pierre Sermanet, et al. 2023. “Video Language Planning.” arXiv. https://doi.org/10.48550/ARXIV.2310.10625.
Du, Yuqing, Ksenia Konyushkova, Misha Denil, Akhil Raju, Jessica Landon, Felix Hill, Nando de Freitas, and Serkan Cabi. 2023. “Vision-Language Models as Success Detectors.” arXiv. https://doi.org/10.48550/ARXIV.2303.07280.
Du, Yuqing, Olivia Watkins, Zihan Wang, Cédric Colas, Trevor Darrell, Pieter Abbeel, Abhishek Gupta, and Jacob Andreas. 2023. “Guiding Pretraining in Reinforcement Learning with Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.06692.
Duan, Haonan, Adam Dziedzic, Nicolas Papernot, and Franziska Boenisch. 2023. “Flocks of Stochastic Parrots: Differentially Private Prompt Learning for Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2305.15594.
Duan, Yan, Xi Chen, Rein Houthooft, John Schulman, and Pieter Abbeel. 2016. “Benchmarking Deep Reinforcement Learning for Continuous Control.” arXiv. https://doi.org/10.48550/ARXIV.1604.06778.
Duch, W., R. Adamczak, and K. Grabczewski. 2001. “A New Methodology of Extraction, Optimization and Application of Crisp and Fuzzy Logical Rules.” IEEE Transactions on Neural Networks 12 (March). https://doi.org/10.1109/72.914524.
Duchi, John C., Michael I. Jordan, and Martin J. Wainwright. 2013. “Local Privacy and Statistical Minimax Rates.” 2013 IEEE 54th Annual Symposium on Foundations of Computer Science, October. https://doi.org/10.1109/focs.2013.53.
Ducoffe, Melanie, and Frederic Precioso. 2015. “QBDC: Query by Dropout Committee for Training Deep Supervised Architecture,” November. http://arxiv.org/abs/1511.06412v2.
Duda, Richard O., Peter E. Hart, and Nils J. Nilsson. 1976. “Subjective Bayesian Methods for Rule-Based Inference Systems.” Proceedings of the June 7-10, 1976, National Computer Conference and Exposition on - AFIPS ’76. https://doi.org/10.1145/1499799.1499948.
Dudoit, Sandrine, Jane Fridlyand, and Terence P Speed. 2002. “Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data.” Journal of the American Statistical Association 97 (March). https://doi.org/10.1198/016214502753479248.
Dumaz, Laure, and Bálint Tóth. 2012. “Marginal Densities of the "True" Self-Repelling Motion.” arXiv. https://doi.org/10.48550/ARXIV.1202.4327.
Dumoulin, Vincent, Ishmael Belghazi, Ben Poole, Olivier Mastropietro, Alex Lamb, Martin Arjovsky, and Aaron Courville. 2016. “Adversarially Learned Inference.” arXiv. https://doi.org/10.48550/ARXIV.1606.00704.
Dumoulin, Vincent, Ethan Perez, Nathan Schucher, Florian Strub, Harm Vries, Aaron Courville, and Yoshua Bengio. 2018. “Feature-Wise Transformations.” Distill 3 (July). https://doi.org/10.23915/distill.00011.
Dumoulin, Vincent, Jonathon Shlens, and Manjunath Kudlur. 2016. “A Learned Representation for Artistic Style.” arXiv. https://doi.org/10.48550/ARXIV.1610.07629.
Dunson, David B., Natesh Pillai, and Ju-Hyun Park. 2007. “Bayesian Density Regression.” Journal of the Royal Statistical Society Series B: Statistical Methodology 69 (March). https://doi.org/10.1111/j.1467-9868.2007.00582.x.
Duong, Manh Khoi. 2019. “4.-8. März 2019.” https://doi.org/10.18420/BTW2019-WS-17.
Duquesne, Thomas, and Jean-François Le Gall. 2004. “Probabilistic and Fractal Aspects of Lévy Trees.” Probability Theory and Related Fields 131 (November). https://doi.org/10.1007/s00440-004-0385-4.
Durante, Zane, Qiuyuan Huang, Naoki Wake, Ran Gong, Jae Sung Park, Bidipta Sarkar, Rohan Taori, et al. 2024. “Agent AI: Surveying the Horizons of Multimodal Interaction.” arXiv. https://doi.org/10.48550/ARXIV.2401.03568.
Duroux, Roxane, and Erwan Scornet. 2016. “Impact of Subsampling and Pruning on Random Forests.” arXiv. https://doi.org/10.48550/ARXIV.1603.04261.
Durugkar, Ishan, Ian Gemp, and Sridhar Mahadevan. 2016. “Generative Multi-Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.01673.
Dutt, Arkopal, Edwin Pednault, Chai Wah Wu, Sarah Sheldon, John Smolin, Lev Bishop, and Isaac L. Chuang. 2023. “Active Learning of Quantum System Hamiltonians Yields Query Advantage.” Physical Review Research 5 (July). https://doi.org/10.1103/physrevresearch.5.033060.
Duvenaud, David, Hannes Nickisch, and Carl Edward Rasmussen. 2011. “Additive Gaussian Processes.” arXiv. https://doi.org/10.48550/ARXIV.1112.4394.
Dyer, Chris, Miguel Ballesteros, Wang Ling, Austin Matthews, and Noah A. Smith. 2015. “Transition-Based Dependency Parsing with Stack Long Short-Term Memory.” arXiv. https://doi.org/10.48550/ARXIV.1505.08075.
Dyer, Chris, Adhiguna Kuncoro, Miguel Ballesteros, and Noah A. Smith. 2016. “Recurrent Neural Network Grammars.” arXiv. https://doi.org/10.48550/ARXIV.1602.07776.
Dymetman, Marc, Guillaume Bouchard, and Simon Carter. 2012. “The OS* Algorithm: A Joint Approach to Exact Optimization and Sampling.” arXiv. https://doi.org/10.48550/ARXIV.1207.0742.
E, Weinan, Jiequn Han, and Arnulf Jentzen. 2017. “Deep Learning-Based Numerical Methods for High-Dimensional Parabolic Partial Differential Equations and Backward Stochastic Differential Equations.” Communications in Mathematics and Statistics 5 (November). https://doi.org/10.1007/s40304-017-0117-6.
Eberle, Andreas. 2015. “Reflection Couplings and Contraction Rates for Diffusions.” Probability Theory and Related Fields 166 (October). https://doi.org/10.1007/s00440-015-0673-1.
Ecoffet, Adrien, Joost Huizinga, Joel Lehman, Kenneth O. Stanley, and Jeff Clune. 2021. “First Return, Then Explore.” Nature 590 (February). https://doi.org/10.1038/s41586-020-03157-9.
Edwards, Harrison, and Amos Storkey. 2015. “Censoring Representations with an Adversary.” arXiv. https://doi.org/10.48550/ARXIV.1511.05897.
———. 2016. “Towards a Neural Statistician.” arXiv. https://doi.org/10.48550/ARXIV.1606.02185.
Efron, Bradley. 2020. “Prediction, Estimation, and Attribution.” International Statistical Review 88 (December). https://doi.org/10.1111/insr.12409.
Egami, Naoki, Christian J. Fong, Justin Grimmer, Margaret E. Roberts, and Brandon M. Stewart. 2018. “How to Make Causal Inferences Using Texts,” February. http://arxiv.org/abs/1802.02163v1.
Ehrlinger, John. 2016. “ggRandomForests: Exploring Random Forest Survival.” arXiv. https://doi.org/10.48550/ARXIV.1612.08974.
Eichler, Michael. 2011. “Graphical Modelling of Multivariate Time Series.” Probability Theory and Related Fields 153 (February). https://doi.org/10.1007/s00440-011-0345-8.
Eigen, David, Christian Puhrsch, and Rob Fergus. 2014. “Depth Map Prediction from a Single Image Using a Multi-Scale Deep Network.” arXiv. https://doi.org/10.48550/ARXIV.1406.2283.
Eigen, David, Marc’Aurelio Ranzato, and Ilya Sutskever. 2013. “Learning Factored Representations in a Deep Mixture of Experts.” arXiv. https://doi.org/10.48550/ARXIV.1312.4314.
Eigen, David, Jason Rolfe, Rob Fergus, and Yann LeCun. 2013. “Understanding Deep Architectures Using a Recursive Convolutional Network.” arXiv. https://doi.org/10.48550/ARXIV.1312.1847.
Eksombatchai, Chantat, Pranav Jindal, Jerry Zitao Liu, Yuchen Liu, Rahul Sharma, Charles Sugnet, Mark Ulrich, and Jure Leskovec. 2017. “Pixie: A System for Recommending 3+ Billion Items to 200+ Million Users in Real-Time,” November. http://arxiv.org/abs/1711.07601v1.
Elben, Andreas, Steven T. Flammia, Hsin-Yuan Huang, Richard Kueng, John Preskill, Benoît Vermersch, and Peter Zoller. 2022. “The Randomized Measurement Toolbox.” Nature Reviews Physics 5 (December). https://doi.org/10.1038/s42254-022-00535-2.
Elgarøy, Ø, and T Multamäki. 2006. “Bayesian Analysis of Friedmannless Cosmologies.” Journal of Cosmology and Astroparticle Physics 2006 (September). https://doi.org/10.1088/1475-7516/2006/09/002.
Elhamifar, Ehsan, Guillermo Sapiro, and S. Shankar Sastry. 2014. “Dissimilarity-Based Sparse Subset Selection.” arXiv. https://doi.org/10.48550/ARXIV.1407.6810.
Elias, Isaac, Heiga Zen, Jonathan Shen, Yu Zhang, Ye Jia, RJ Skerry-Ryan, and Yonghui Wu. 2021. “Parallel Tacotron 2: A Non-Autoregressive Neural TTS Model with Differentiable Duration Modeling.” arXiv. https://doi.org/10.48550/ARXIV.2103.14574.
Elmi, Sayda, and Kian-Lee Tan. 2021. “DeepFEC: Energy Consumption Prediction Under Real-World Driving Conditions for Smart Cities.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3449983.
Elmougy, Youssef, and Ling Liu. 2023. “Demystifying Fraudulent Transactions and Illicit Nodes in the Bitcoin Network for Financial Forensics.” Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3580305.3599803.
Elnaggar, Ahmed, Hazem Essam, Wafaa Salah-Eldin, Walid Moustafa, Mohamed Elkerdawy, Charlotte Rochereau, and Burkhard Rost. 2023. “Ankh ☥: Optimized Protein Language Model Unlocks General-Purpose Modelling,” January. https://doi.org/10.1101/2023.01.16.524265.
Elshawi, Radwa, Mohamed Maher, and Sherif Sakr. 2019. “Automated Machine Learning: State-of-the-Art and Open Challenges.” arXiv. https://doi.org/10.48550/ARXIV.1906.02287.
El-Shishtawy, T., and A. El-Sammak. 2014. “The Best Templates Match Technique for Example Based Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1406.1241.
Elsken, Thomas, Jan-Hendrik Metzen, and Frank Hutter. 2017. “Simple and Efficient Architecture Search for Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1711.04528.
Elthakeb, Ahmed T., Prannoy Pilligundla, FatemehSadat Mireshghallah, Amir Yazdanbakhsh, and Hadi Esmaeilzadeh. 2018. “ReLeQ: A Reinforcement Learning Approach for Deep Quantization of Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1811.01704.
Elzen, Stef van den, and Jarke J. van Wijk. 2011. “BaobabView: Interactive Construction and Analysis of Decision Trees.” 2011 IEEE Conference on Visual Analytics Science and Technology (VAST), October. https://doi.org/10.1109/vast.2011.6102453.
“Encyclopedia of Biostatistics.” 2005, February. https://doi.org/10.1002/0470011815.
“Encyclopedia of Machine Learning and Data Mining.” 2017. https://doi.org/10.1007/978-1-4899-7687-1.
Engel, Jesse, Cinjon Resnick, Adam Roberts, Sander Dieleman, Douglas Eck, Karen Simonyan, and Mohammad Norouzi. 2017. “Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders.” arXiv. https://doi.org/10.48550/ARXIV.1704.01279.
Engel, Nico, Vasileios Belagiannis, and Klaus Dietmayer. 2021. “Point Transformer.” IEEE Access 9. https://doi.org/10.1109/access.2021.3116304.
Erdős, László, Horng-Tzer Yau, and Jun Yin. 2011. “Bulk Universality for Generalized Wigner Matrices.” Probability Theory and Related Fields 154 (October). https://doi.org/10.1007/s00440-011-0390-3.
Ericsson, Linus, Henry Gouk, Chen Change Loy, and Timothy M. Hospedales. 2022. “Self-Supervised Representation Learning: Introduction, Advances, and Challenges.” IEEE Signal Processing Magazine 39 (May). https://doi.org/10.1109/msp.2021.3134634.
Eriguchi, Akiko, Yoshimasa Tsuruoka, and Kyunghyun Cho. 2017. “Learning to Parse and Translate Improves Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1702.03525.
Erven, Tim van, and Botond Szabó. 2021. “Fast Exact Bayesian Inference for Sparse Signals in the Normal Sequence Model.” Bayesian Analysis 16 (September). https://doi.org/10.1214/20-ba1227.
Escalante, Hugo Jair. 2020. “Automated Machine Learning – a Brief Review at the End of the Early Years.” arXiv. https://doi.org/10.48550/ARXIV.2008.08516.
Escobar, Michael D., and Mike West. 1995. “Bayesian Density Estimation and Inference Using Mixtures.” Journal of the American Statistical Association 90 (June). https://doi.org/10.1080/01621459.1995.10476550.
“ESPnet2 Pretrained Model, Kamo-Naoyuki/Librispeech_asr_train_asr_conformer6_n_fft512_hop_length256_raw_en_bpe5000_scheduler_confwarmup_steps40000_optim_conflr0.0025_sp_valid.acc.ave, Fs=16k, Lang=en.” 2021, March. https://doi.org/10.5281/ZENODO.4604066.
Esteban, Cristóbal, Stephanie L. Hyland, and Gunnar Rätsch. 2017. “Real-Valued (Medical) Time Series Generation with Recurrent Conditional GANs.” arXiv. https://doi.org/10.48550/ARXIV.1706.02633.
“Experiments with the Graph Traverser Program.” 1966. Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences 294 (September). https://doi.org/10.1098/rspa.1966.0205.
Eyben, Florian. 2016. “Real-Time Speech and Music Classification by Large Audio Feature Space Extraction.” Springer Theses. https://doi.org/10.1007/978-3-319-27299-3.
Faal, Farshid, Ketra Schmitt, and Jia Yuan Yu. 2022. “Reward Modeling for Mitigating Toxicity in Transformer-Based Language Models.” Applied Intelligence 53 (July). https://doi.org/10.1007/s10489-022-03944-z.
Faber, Felix A., Luke Hutchison, Bing Huang, Justin Gilmer, Samuel S. Schoenholz, George E. Dahl, Oriol Vinyals, Steven Kearnes, Patrick F. Riley, and O. Anatole von Lilienfeld. 2017. “Prediction Errors of Molecular Machine Learning Models Lower Than Hybrid DFT Error.” Journal of Chemical Theory and Computation 13 (October). https://doi.org/10.1021/acs.jctc.7b00577.
Fader, Anthony, Luke Zettlemoyer, and Oren Etzioni. 2014. “Open Question Answering over Curated and Extracted Knowledge Bases.” Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2623330.2623677.
Falcon, William, and Kyunghyun Cho. 2020. “A Framework for Contrastive Self-Supervised Learning and Designing a New Approach.” arXiv. https://doi.org/10.48550/ARXIV.2009.00104.
Faletto, Gregory. 2023. “Fused Extended Two-Way Fixed Effects for Difference-in-Differences with Staggered Adoptions.” arXiv. https://doi.org/10.48550/ARXIV.2312.05985.
Faltejsková, Kateřina, and Jiří Vondrášek. 2023. “PAPerFly: Partial Assembly-Based Peak Finder for Ab Initio Binding Site Reconstruction.” BMC Bioinformatics 24 (December). https://doi.org/10.1186/s12859-023-05613-5.
Fan, Angela, Edouard Grave, and Armand Joulin. 2019. “Reducing Transformer Depth on Demand with Structured Dropout.” arXiv. https://doi.org/10.48550/ARXIV.1909.11556.
Fan, Guofan, Zekun Qi, Wenkai Shi, and Kaisheng Ma. 2023. “Point-GCC: Universal Self-Supervised 3D Scene Pre-Training via Geometry-Color Contrast.” arXiv. https://doi.org/10.48550/ARXIV.2305.19623.
Fan, Jianqing, and Jinchi Lv. 2018. “Sure Independence Screening.” Wiley StatsRef: Statistics Reference Online, May. https://doi.org/10.1002/9781118445112.stat08043.
Fan, Jianqing, and Wenyang Zhang. 1999. “Statistical Estimation in Varying Coefficient Models.” The Annals of Statistics 27 (October). https://doi.org/10.1214/aos/1017939139.
Fan, Lizhou, Lingyao Li, Zihui Ma, Sanggyu Lee, Huizi Yu, and Libby Hemphill. 2023. “A Bibliometric Review of Large Language Models Research from 2017 to 2023.” arXiv. https://doi.org/10.48550/ARXIV.2304.02020.
Fan, Miao, Jiacheng Guo, Shuai Zhu, Shuo Miao, Mingming Sun, and Ping Li. 2019. “MOBIUS.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330651.
Fan, Shaohua, Chuan Shi, and Xiao Wang. 2018. “Abnormal Event Detection via Heterogeneous Information Network Embedding.” Proceedings of the 27th ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3269206.3269281.
Fan, Wenqi, Yao Ma, Qing Li, Yuan He, Eric Zhao, Jiliang Tang, and Dawei Yin. 2019. “Graph Neural Networks for Social Recommendation.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313488.
Fang, Han, Pengfei Xiong, Luhui Xu, and Yu Chen. 2021. “CLIP2Video: Mastering Video-Text Retrieval via Image CLIP.” arXiv. https://doi.org/10.48550/ARXIV.2106.11097.
Fang, Hao, Saurabh Gupta, Forrest Iandola, Rupesh Srivastava, Li Deng, Piotr Dollár, Jianfeng Gao, et al. 2014. “From Captions to Visual Concepts and Back.” arXiv. https://doi.org/10.48550/ARXIV.1411.4952.
Fang, Tongtong, Nan Lu, Gang Niu, and Masashi Sugiyama. 2020. “Rethinking Importance Weighting for Deep Learning Under Distribution Shift.” arXiv. https://doi.org/10.48550/ARXIV.2006.04662.
Fanourakis, Nikolaos, Vasilis Efthymiou, Dimitris Kotzinos, and Vassilis Christophides. 2023. “Knowledge Graph Embedding Methods for Entity Alignment: Experimental Review.” Data Mining and Knowledge Discovery 37 (June). https://doi.org/10.1007/s10618-023-00941-9.
Farnia, Farzan, and Asuman Ozdaglar. 2020. “GANs May Have No Nash Equilibria.” arXiv. https://doi.org/10.48550/ARXIV.2002.09124.
Faruqui, Manaal, Yulia Tsvetkov, Dani Yogatama, Chris Dyer, and Noah Smith. 2015. “Sparse Overcomplete Word Vector Representations.” arXiv. https://doi.org/10.48550/ARXIV.1506.02004.
Faul, Anita C., and Michael E. Tipping. 2002. “Analysis of Sparse Bayesian Learning.” Advances in Neural Information Processing Systems 14, November. https://doi.org/10.7551/mitpress/1120.003.0054.
Favaro, S., A. Lijoi, C. Nava, B. Nipoti, I. Prünster, and Y. W. Teh. 2016. “On the Stick-Breaking Representation for Homogeneous NRMIs.” Bayesian Analysis 11 (September). https://doi.org/10.1214/15-ba964.
Feder, Amir, Nadav Oved, Uri Shalit, and Roi Reichart. 2021. “CausaLM: Causal Model Explanation Through Counterfactual Language Models.” Computational Linguistics, May. https://doi.org/10.1162/coli_a_00404.
Fedus, William, Barret Zoph, and Noam Shazeer. 2021. “Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity.” arXiv. https://doi.org/10.48550/ARXIV.2101.03961.
Feldman, Jacob. 1997. “The Structure of Perceptual Categories.” Journal of Mathematical Psychology 41 (June). https://doi.org/10.1006/jmps.1997.1154.
Fellegi, Ivan P., and Alan B. Sunter. 1969. “A Theory for Record Linkage.” Journal of the American Statistical Association 64 (December). https://doi.org/10.1080/01621459.1969.10501049.
Feng, Aosong, Chenyu You, Shiqiang Wang, and Leandros Tassiulas. 2022. “KerGNNs: Interpretable Graph Neural Networks with Graph Kernels.” arXiv. https://doi.org/10.48550/ARXIV.2201.00491.
Feng, Felicia Li, Ryan Yen, Yuzhe You, Mingming Fan, Jian Zhao, and Zhicong Lu. 2023. “CoPrompt: Supporting Prompt Sharing and Referring in Collaborative Natural Language Programming.” arXiv. https://doi.org/10.48550/ARXIV.2310.09235.
Feng, Fuli, Weiran Huang, Xiangnan He, Xin Xin, Qifan Wang, and Tat-Seng Chua. 2021. “Should Graph Convolution Trust Neighbors? A Simple Causal Inference Method.” Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3404835.3462971.
Feng, Jiashi, Huan Xu, and Shie Mannor. 2017. “Outlier Robust Online Learning.” arXiv. https://doi.org/10.48550/ARXIV.1701.00251.
Feng, Minwei, Bing Xiang, Michael R. Glass, Lidan Wang, and Bowen Zhou. 2015. “Applying Deep Learning to Answer Selection: A Study and an Open Task.” arXiv. https://doi.org/10.48550/ARXIV.1508.01585.
Feng, Rui, Yang Yang, Wenjie Hu, Fei Wu, and Yueting Zhuang. 2017. “Representation Learning for Scale-Free Networks.” arXiv. https://doi.org/10.48550/ARXIV.1711.10755.
Feng, Tzu-hsun, Annie Dong, Ching-Feng Yeh, Shu-wen Yang, Tzu-Quan Lin, Jiatong Shi, Kai-Wei Chang, et al. 2022. “SUPERB @ SLT 2022: Challenge on Generalization and Efficiency of Self-Supervised Speech Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.2210.08634.
Feng, Xiaocheng, Lifu Huang, Duyu Tang, Heng Ji, Bing Qin, and Ting Liu. 2016. “A Language-Independent Neural Network for Event Detection.” Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). https://doi.org/10.18653/v1/p16-2011.
Feng, Yifan, Haoxuan You, Zizhao Zhang, Rongrong Ji, and Yue Gao. 2018. “Hypergraph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1809.09401.
Feng, Yufei, Binbin Hu, Fuyu Lv, Qingwen Liu, Zhiqiang Zhang, and Wenwu Ou. 2020. “ATBRG: Adaptive Target-Behavior Relational Graph Network for Effective Recommendation.” Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3397271.3401428.
Feng, Zhida, Zhenyu Zhang, Xintong Yu, Yewei Fang, Lanxin Li, Xuyi Chen, Yuxiang Lu, et al. 2023. “ERNIE-ViLG 2.0: Improving Text-to-Image Diffusion Model with Knowledge-Enhanced Mixture-of-Denoising-Experts.” 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52729.2023.00977.
Fergus, Suzanne, Michelle Botha, and Mehrnoosh Ostovar. 2023. “Evaluating Academic Answers Generated Using ChatGPT.” Journal of Chemical Education 100 (March). https://doi.org/10.1021/acs.jchemed.3c00087.
Fernando, Chrisantha, Dylan Banarse, Charles Blundell, Yori Zwols, David Ha, Andrei A. Rusu, Alexander Pritzel, and Daan Wierstra. 2017. “PathNet: Evolution Channels Gradient Descent in Super Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1701.08734.
Ferov, Michal, and Marek Modrý. 2016. “Enhancing LambdaMART Using Oblivious Trees.” arXiv. https://doi.org/10.48550/ARXIV.1609.05610.
Ferrara, Emilio. 2023. “Should ChatGPT Be Biased? Challenges and Risks of Bias in Large Language Models.” First Monday, November. https://doi.org/10.5210/fm.v28i11.13346.
Fetahu, Besnik, Avishek Anand, and Maria Koutraki. 2019. “TableNet: An Approach for Determining Fine-Grained Relations for Wikipedia Tables.” arXiv. https://doi.org/10.48550/ARXIV.1902.01740.
Fey, Matthias, and Jan Eric Lenssen. 2019. “Fast Graph Representation Learning with PyTorch Geometric.” arXiv. https://doi.org/10.48550/ARXIV.1903.02428.
Filos, Angelos, Clare Lyle, Yarin Gal, Sergey Levine, Natasha Jaques, and Gregory Farquhar. 2021. “PsiPhi-Learning: Reinforcement Learning with Demonstrations Using Successor Features and Inverse Temporal Difference Learning.” arXiv. https://doi.org/10.48550/ARXIV.2102.12560.
Finkel, Jenny Rose, Trond Grenager, and Christopher Manning. 2005. “Incorporating Non-Local Information into Information Extraction Systems by Gibbs Sampling.” Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics - ACL ’05. https://doi.org/10.3115/1219840.1219885.
Finkelstein, Alexander, Uri Almog, and Mark Grobman. 2019. “Fighting Quantization Bias with Bias.” arXiv. https://doi.org/10.48550/ARXIV.1906.03193.
Finkenstädt, Bärbel, Dan J. Woodcock, Michal Komorowski, Claire V. Harper, Julian R. E. Davis, Mike R. H. White, and David A. Rand. 2013. “Quantifying Intrinsic and Extrinsic Noise in Gene Transcription Using the Linear Noise Approximation: An Application to Single Cell Data.” The Annals of Applied Statistics 7 (December). https://doi.org/10.1214/13-aoas669.
Finn, Chelsea, and Sergey Levine. 2017. “Meta-Learning and Universality: Deep Representations and Gradient Descent Can Approximate Any Learning Algorithm.” arXiv. https://doi.org/10.48550/ARXIV.1710.11622.
Firat, Orhan, Baskaran Sankaran, Yaser Al-Onaizan, Fatos T. Yarman Vural, and Kyunghyun Cho. 2016. “Zero-Resource Translation with Multi-Lingual Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1606.04164.
Flandoli, Franco, and Dariusz Gatarek. 1995. “Martingale and Stationary Solutions for Stochastic Navier-Stokes Equations.” Probability Theory and Related Fields 102 (September). https://doi.org/10.1007/bf01192467.
Floreano, Dario, Peter Dürr, and Claudio Mattiussi. 2008. “Neuroevolution: From Architectures to Learning.” Evolutionary Intelligence 1 (January). https://doi.org/10.1007/s12065-007-0002-4.
Foerster, Jakob N., Yannis M. Assael, Nando de Freitas, and Shimon Whiteson. 2016. “Learning to Communicate with Deep Multi-Agent Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1605.06676.
Foerster, Jakob, Nantas Nardelli, Gregory Farquhar, Triantafyllos Afouras, Philip H. S. Torr, Pushmeet Kohli, and Shimon Whiteson. 2017. “Stabilising Experience Replay for Deep Multi-Agent Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1702.08887.
Föllmer, Hans, and D. Kramkov. 2006. “Optional Decompositions Under Constraints,” June. https://doi.org/10.18452/3814.
Fop, Michael, Keith M. Smart, and Thomas Brendan Murphy. 2017. “Variable Selection for Latent Class Analysis with Application to Low Back Pain Diagnosis.” The Annals of Applied Statistics 11 (December). https://doi.org/10.1214/17-aoas1061.
Foresee, F. Dan, and M. T. Hagan. n.d. “Gauss-Newton Approximation to Bayesian Learning.” Proceedings of International Conference on Neural Networks (ICNN’97). https://doi.org/10.1109/icnn.1997.614194.
Fortuin, Vincent, Matthias Hüser, Francesco Locatello, Heiko Strathmann, and Gunnar Rätsch. 2018. “SOM-VAE: Interpretable Discrete Representation Learning on Time Series.” arXiv. https://doi.org/10.48550/ARXIV.1806.02199.
Fortunato, Meire, Mohammad Gheshlaghi Azar, Bilal Piot, Jacob Menick, Ian Osband, Alex Graves, Vlad Mnih, et al. 2017. “Noisy Networks for Exploration.” arXiv. https://doi.org/10.48550/ARXIV.1706.10295.
Fortunato, Meire, Charles Blundell, and Oriol Vinyals. 2017. “Bayesian Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1704.02798.
Foster, Adam, Martin Jankowiak, Matthew O’Meara, Yee Whye Teh, and Tom Rainforth. 2019. “A Unified Stochastic Gradient Approach to Designing Bayesian-Optimal Experiments.” arXiv. https://doi.org/10.48550/ARXIV.1911.00294.
Foster, Dylan J., Sham M. Kakade, Jian Qian, and Alexander Rakhlin. 2021. “The Statistical Complexity of Interactive Decision Making.” arXiv. https://doi.org/10.48550/ARXIV.2112.13487.
Foti, Nicholas J., Jason Xu, Dillon Laird, and Emily B. Fox. 2014. “Stochastic Variational Inference for Hidden Markov Models.” arXiv. https://doi.org/10.48550/ARXIV.1411.1670.
Fournier, Nicolas, and Arnaud Guillin. 2014. “On the Rate of Convergence in Wasserstein Distance of the Empirical Measure.” Probability Theory and Related Fields 162 (October). https://doi.org/10.1007/s00440-014-0583-7.
“Fourteenth ACM Conference on Recommender Systems.” 2020, September. https://doi.org/10.1145/3383313.
Fowkes, Jaroslav, and Charles Sutton. 2016. “A Subsequence Interleaving Model for Sequential Pattern Mining.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939787.
Fox, Emily B., Erik B. Sudderth, Michael I. Jordan, and Alan S. Willsky. 2011. “A Sticky HDP-HMM with Application to Speaker Diarization.” The Annals of Applied Statistics 5 (June). https://doi.org/10.1214/10-aoas395.
Fox, James, and Sivasankaran Rajamanickam. 2019. “How Robust Are Graph Neural Networks to Structural Noise?” arXiv. https://doi.org/10.48550/ARXIV.1912.10206.
Fradi, A., C. Samir, J. Braga, S. H. Joshi, and J-M. Loubes. 2022. “Nonparametric Bayesian Regression and Classification on Manifolds, with Applications to 3D Cochlear Shapes.” IEEE Transactions on Image Processing 31. https://doi.org/10.1109/tip.2022.3147971.
Frankle, Jonathan, and David Bau. 2019. “Dissecting Pruned Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1907.00262.
Frankle, Jonathan, and Michael Carbin. 2018. “The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1803.03635.
Frantar, Elias, and Dan Alistarh. 2023. “SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot.” arXiv. https://doi.org/10.48550/ARXIV.2301.00774.
Frantar, Elias, Saleh Ashkboos, Torsten Hoefler, and Dan Alistarh. 2022. “GPTQ: Accurate Post-Training Quantization for Generative Pre-Trained Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2210.17323.
Fraser, Kathleen C., Jed A. Meltzer, Naida L. Graham, Carol Leonard, Graeme Hirst, Sandra E. Black, and Elizabeth Rochon. 2014. “Automated Classification of Primary Progressive Aphasia Subtypes from Narrative Speech Transcripts.” Cortex 55 (June). https://doi.org/10.1016/j.cortex.2012.12.006.
Frazee, Alyssa C, Geo Pertea, Andrew E Jaffe, Ben Langmead, Steven L Salzberg, and Jeffrey T Leek. 2014. “Flexible Analysis of Transcriptome Assemblies with Ballgown,” March. https://doi.org/10.1101/003665.
Friedman, Jerome, Trevor Hastie, and Robert Tibshirani. 2000. “Additive Logistic Regression: A Statistical View of Boosting (with Discussion and a Rejoinder by the Authors).” The Annals of Statistics 28 (April). https://doi.org/10.1214/aos/1016218223.
Friedman, Nir. 2013. “The Bayesian Structural EM Algorithm.” arXiv. https://doi.org/10.48550/ARXIV.1301.7373.
Friedrich, Felix, Katharina Hämmerl, Patrick Schramowski, Jindrich Libovicky, Kristian Kersting, and Alexander Fraser. 2024. “Multilingual Text-to-Image Generation Magnifies Gender Stereotypes and Prompt Engineering May Not Help You.” arXiv. https://doi.org/10.48550/ARXIV.2401.16092.
Friesen, Abram L., and Pedro Domingos. 2016. “Recursive Decomposition for Nonconvex Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1611.02755.
Fritz, Cornelius, Marius Mehrl, Paul W. Thurner, and Göran Kauermann. 2021. “All That Glitters Is Not Gold: Relational Events Models with Spurious Events.” arXiv. https://doi.org/10.48550/ARXIV.2109.10348.
Frühwirth-Schnatter, Sylvia, Gertraud Malsiner-Walli, and Bettina Grün. 2021. “Generalized Mixtures of Finite Mixtures and Telescoping Sampling.” Bayesian Analysis 16 (December). https://doi.org/10.1214/21-ba1294.
Fu, Chaoyou, Peixian Chen, Yunhang Shen, Yulei Qin, Mengdan Zhang, Xu Lin, Jinrui Yang, et al. 2023. “MME: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2306.13394.
Fu, Cheng, Shilin Zhu, Hao Su, Ching-En Lee, and Jishen Zhao. 2018. “Towards Fast and Energy-Efficient Binarized Neural Network Inference on FPGA.” arXiv. https://doi.org/10.48550/ARXIV.1810.02068.
Fu, Hao, Chunyuan Li, Xiaodong Liu, Jianfeng Gao, Asli Celikyilmaz, and Lawrence Carin. 2019. “Cyclical Annealing Schedule: A Simple Approach to Mitigating KL Vanishing.” arXiv. https://doi.org/10.48550/ARXIV.1903.10145.
Fu, Jinlan, See-Kiong Ng, and Pengfei Liu. 2022. “Polyglot Prompt: Multilingual Multitask PrompTraining.” arXiv. https://doi.org/10.48550/ARXIV.2204.14264.
Fu, Qian-Jie, Sherol Chinchilla, and John J. Galvin. 2004. “The Role of Spectral and Temporal Cues in Voice Gender Discrimination by Normal-Hearing Listeners and Cochlear Implant Users.” Journal of the Association for Research in Otolaryngology 5 (May). https://doi.org/10.1007/s10162-004-4046-1.
Fujita, Yuya, Shinji Watanabe, Motoi Omachi, and Xuankai Chan. 2020. “Insertion-Based Modeling for End-to-End Automatic Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2005.13211.
Furelos-Blanco, Daniel, Mark Law, Anders Jonsson, Krysia Broda, and Alessandra Russo. 2022. “Hierarchies of Reward Machines.” arXiv. https://doi.org/10.48550/ARXIV.2205.15752.
G. Matthews, Alexander G. de, Mark van der Wilk, Tom Nickson, Keisuke Fujii, Alexis Boukouvalas, Pablo León-Villagrá, Zoubin Ghahramani, and James Hensman. 2016. “GPflow: A Gaussian Process Library Using TensorFlow.” arXiv. https://doi.org/10.48550/ARXIV.1610.08733.
Gabriel, Iason. 2020. “Artificial Intelligence, Values, and Alignment.” Minds and Machines 30 (September). https://doi.org/10.1007/s11023-020-09539-2.
———. 2022. “Toward a Theory of Justice for Artificial Intelligence.” Daedalus 151. https://doi.org/10.1162/daed_a_01911.
Gal, Rinon, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, and Daniel Cohen-Or. 2022. “An Image Is Worth One Word: Personalizing Text-to-Image Generation Using Textual Inversion.” arXiv. https://doi.org/10.48550/ARXIV.2208.01618.
Gal, Yarin, and Zoubin Ghahramani. 2015. “Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1506.02142.
Galárraga, Luis, Simon Razniewski, Antoine Amarilli, and Fabian M. Suchanek. 2017. “Predicting Completeness in Knowledge Bases.” Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/3018661.3018739.
Gama, João, Ricardo Rocha, and Pedro Medas. 2003. “Accurate Decision Trees for Mining High-Speed Data Streams.” Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/956750.956813.
Gamboa, John Cristian Borges. 2017. “Deep Learning for Time-Series Analysis.” arXiv. https://doi.org/10.48550/ARXIV.1701.01887.
Gan, Zhe, Yunchen Pu, Ricardo Henao, Chunyuan Li, Xiaodong He, and Lawrence Carin. 2016. “Learning Generic Sentence Representations Using Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.07897.
Gandikota, Rohit, Joanna Materzynska, Jaden Fiotto-Kaufman, and David Bau. 2023. “Erasing Concepts from Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.07345.
Ganea, Octavian-Eugen, Gary Bécigneul, and Thomas Hofmann. 2018. “Hyperbolic Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1805.09112.
Ganin, Yaroslav, and Victor Lempitsky. 2014. “Unsupervised Domain Adaptation by Backpropagation.” arXiv. https://doi.org/10.48550/ARXIV.1409.7495.
Gao, Chen, Yu Zheng, Wenjie Wang, Fuli Feng, Xiangnan He, and Yong Li. 2024. “Causal Inference in Recommender Systems: A Survey and Future Directions.” ACM Transactions on Information Systems, January. https://doi.org/10.1145/3639048.
Gao, G., M. Zafari, and A. C. Reynolds. 2005. “Quantifying Uncertainty for the PUNQ-S3 Problem in a Bayesian Setting with RML and EnKF.” All Days, January. https://doi.org/10.2118/93324-ms.
Gao, Hongyang, Yongjun Chen, and Shuiwang Ji. 2019. “Learning Graph Pooling and Hybrid Convolutional Operations for Text Representations.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313395.
Gao, Hongyang, Zhengyang Wang, and Shuiwang Ji. 2018. “Large-Scale Learnable Graph Convolutional Networks.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3219947.
Gao, Jingkun, Xiaomin Song, Qingsong Wen, Pichao Wang, Liang Sun, and Huan Xu. 2020. “RobustTAD: Robust Time Series Anomaly Detection via Decomposition and Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2002.09545.
Gao, Jun, Chengcheng Tang, Vignesh Ganapathi-Subramanian, Jiahui Huang, Hao Su, and Leonidas J. Guibas. 2019. “DeepSpline: Data-Driven Reconstruction of Parametric Curves and Surfaces.” arXiv. https://doi.org/10.48550/ARXIV.1901.03781.
Gao, Leo, Stella Biderman, Sid Black, Laurence Golding, Travis Hoppe, Charles Foster, Jason Phang, et al. 2021. “The Pile: An 800GB Dataset of Diverse Text for Language Modeling.” arXiv. https://doi.org/10.48550/ARXIV.2101.00027.
Gao, Mingfei, Ang Li, Ruichi Yu, Vlad I. Morariu, and Larry S. Davis. 2017. “C-WSL: Count-Guided Weakly Supervised Localization.” arXiv. https://doi.org/10.48550/ARXIV.1711.05282.
Gao, Peng, Jiaming Han, Renrui Zhang, Ziyi Lin, Shijie Geng, Aojun Zhou, Wei Zhang, et al. 2023. “LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model.” arXiv. https://doi.org/10.48550/ARXIV.2304.15010.
Gao, Qiaozi, Govind Thattai, Suhaila Shakiah, Xiaofeng Gao, Shreyas Pansare, Vasu Sharma, Gaurav Sukhatme, et al. 2023. “Alexa Arena: A User-Centric Interactive Platform for Embodied AI.” arXiv. https://doi.org/10.48550/ARXIV.2303.01586.
Gao, Ruohan, Zilin Si, Yen-Yu Chang, Samuel Clarke, Jeannette Bohg, Li Fei-Fei, Wenzhen Yuan, and Jiajun Wu. 2022. “ObjectFolder 2.0: A Multisensory Object Dataset for Sim2Real Transfer.” arXiv. https://doi.org/10.48550/ARXIV.2204.02389.
Gao, Weihao, Xiangjun Fan, Chong Wang, Jiankai Sun, Kai Jia, Wenzhi Xiao, Ruofan Ding, Xingyan Bin, Hui Yang, and Xiaobing Liu. 2020. “Deep Retrieval: Learning a Retrievable Structure for Large-Scale Recommendations.” arXiv. https://doi.org/10.48550/ARXIV.2007.07203.
Gao, Yeqi, Zhao Song, and Junze Yin. 2023. “An Iterative Algorithm for Rescaled Hyperbolic Functions Regression.” arXiv. https://doi.org/10.48550/ARXIV.2305.00660.
Gao, Yixin, Yuriy Mishchenko, Anish Shah, Spyros Matsoukas, and Shiv Vitaladevuni. 2020. “Towards Data-Efficient Modeling for Wake Word Spotting.” arXiv. https://doi.org/10.48550/ARXIV.2010.06659.
Gao, Yuan, Xiang Wang, Xiangnan He, Zhenguang Liu, Huamin Feng, and Yongdong Zhang. 2023. “Alleviating Structural Distribution Shift in Graph Anomaly Detection.” Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/3539597.3570377.
Gao, Yunfan, Tao Sheng, Youlin Xiang, Yun Xiong, Haofen Wang, and Jiawei Zhang. 2023. “Chat-REC: Towards Interactive and Explainable LLMs-Augmented Recommender System.” arXiv. https://doi.org/10.48550/ARXIV.2303.14524.
Garbin, Stephan J., Marek Kowalski, Matthew Johnson, Jamie Shotton, and Julien Valentin. 2021. “FastNeRF: High-Fidelity Neural Rendering at 200FPS.” arXiv. https://doi.org/10.48550/ARXIV.2103.10380.
Garcia, Victor, and Joan Bruna. 2017. “Few-Shot Learning with Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1711.04043.
Garcia-Duran, Alberto, and Mathias Niepert. 2017. “Learning Graph Representations with Embedding Propagation.” arXiv. https://doi.org/10.48550/ARXIV.1710.03059.
Garipov, Timur, Dmitry Podoprikhin, Alexander Novikov, and Dmitry Vetrov. 2016. “Ultimate Tensorization: Compressing Convolutional and FC Layers Alike.” arXiv. https://doi.org/10.48550/ARXIV.1611.03214.
Garofalakis, Minos, Dongjoon Hyun, Rajeev Rastogi, and Kyuseok Shim. 2000. “Efficient Algorithms for Constructing Decision Trees with Constraints.” Proceedings of the Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/347090.347163.
Gatterbauer, Wolfgang. 2015. “The Linearization of Belief Propagation on Pairwise Markov Networks.” arXiv. https://doi.org/10.48550/ARXIV.1502.04956.
Gatys, Leon A., Alexander S. Ecker, and Matthias Bethge. 2015. “Texture Synthesis Using Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1505.07376.
Gautam, Chandan, Ramesh Balaji, K Sudharsan, Aruna Tiwari, and Kapil Ahuja. 2018. “Localized Multiple Kernel Learning for Anomaly Detection: One-Class Classification.” arXiv. https://doi.org/10.48550/ARXIV.1805.07892.
Gavrikov, Paul, Janis Keuper, and Margret Keuper. 2023. “An Extended Study of Human-Like Behavior Under Adversarial Training.” arXiv. https://doi.org/10.48550/ARXIV.2303.12669.
Gavrilev, Dmitrii, and Evgeny Burnaev. 2023. “Anomaly Detection in Networks via Score-Based Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.2306.15324.
Gawlikowski, Jakob, Cedrique Rovile Njieutcheu Tassi, Mohsin Ali, Jongseok Lee, Matthias Humt, Jianxiang Feng, Anna Kruspe, et al. 2021. “A Survey of Uncertainty in Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2107.03342.
Ge, Changhan, Zihui Ge, Xuan Liu, Ajay Mahimkar, Yusef Shaqalle, Yu Xiang, and Shomik Pathak. 2023. “Chroma: Learning and Using Network Contexts to Reinforce Performance Improving Configurations.” Proceedings of the 29th Annual International Conference on Mobile Computing and Networking, October. https://doi.org/10.1145/3570361.3613256.
Ge, Congcong, Xiaoze Liu, Lu Chen, Yunjun Gao, and Baihua Zheng. 2021. “LargeEA.” Proceedings of the VLDB Endowment 15 (October). https://doi.org/10.14778/3489496.3489504.
Ge, Jiaxin, Hongyin Luo, Siyuan Qian, Yulu Gan, Jie Fu, and Shanghang Zhang. 2023. “Chain of Thought Prompt Tuning in Vision Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.07919.
Ge, Suyu, Lu Cheng, and Huan Liu. 2021. “Improving Cyberbullying Detection with User Interaction.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3449828.
Ge, Tiezheng, Kaiming He, Qifa Ke, and Jian Sun. 2014. “Optimized Product Quantization.” IEEE Transactions on Pattern Analysis and Machine Intelligence 36 (April). https://doi.org/10.1109/tpami.2013.240.
Ge, ZongYuan, Sergey Demyanov, Zetao Chen, and Rahil Garnavi. 2017. “Generative OpenMax for Multi-Class Open Set Classification.” arXiv. https://doi.org/10.48550/ARXIV.1707.07418.
Gehring, Jonas, Michael Auli, David Grangier, and Yann N. Dauphin. 2016. “A Convolutional Encoder Model for Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1611.02344.
Gehring, Jonas, Michael Auli, David Grangier, Denis Yarats, and Yann N. Dauphin. 2017. “Convolutional Sequence to Sequence Learning.” arXiv. https://doi.org/10.48550/ARXIV.1705.03122.
Gelman, A., W. R. Gilks, and G. O. Roberts. 1997. “Weak Convergence and Optimal Scaling of Random Walk Metropolis Algorithms.” The Annals of Applied Probability 7 (February). https://doi.org/10.1214/aoap/1034625254.
Gelman, Andrew, Jessica Hwang, and Aki Vehtari. 2013. “Understanding Predictive Information Criteria for Bayesian Models.” arXiv. https://doi.org/10.48550/ARXIV.1307.5928.
Gelman, Andrew, Daniel Lee, and Jiqiang Guo. 2015. “Stan.” Journal of Educational and Behavioral Statistics 40 (October). https://doi.org/10.3102/1076998615606113.
Gelman, Andrew, and Xiao-Li Meng. 1998. “Simulating Normalizing Constants: From Importance Sampling to Bridge Sampling to Path Sampling.” Statistical Science 13 (May). https://doi.org/10.1214/ss/1028905934.
Gelman, Andrew, and Donald B Rubin. 1996. “Markov Chain Monte Carlo Methods in Biostatistics.” Statistical Methods in Medical Research 5 (December). https://doi.org/10.1177/096228029600500402.
Gelman, Andrew, and Aki Vehtari. 2020. “What Are the Most Important Statistical Ideas of the Past 50 Years?” arXiv. https://doi.org/10.48550/ARXIV.2012.00174.
Geman, Stuart, and Donald Geman. 1984. “Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images.” IEEE Transactions on Pattern Analysis and Machine Intelligence PAMI-6 (November). https://doi.org/10.1109/tpami.1984.4767596.
Geng, Shijie, Shuchang Liu, Zuohui Fu, Yingqiang Ge, and Yongfeng Zhang. 2022. “Recommendation as Language Processing (RLP): A Unified Pretrain, Personalized Prompt &Amp; Predict Paradigm (P5).” Proceedings of the 16th ACM Conference on Recommender Systems, September. https://doi.org/10.1145/3523227.3546767.
Geng, Shijie, Jianbo Yuan, Yu Tian, Yuxiao Chen, and Yongfeng Zhang. 2023. “HiCLIP: Contrastive Language-Image Pretraining with Hierarchy-Aware Attention.” arXiv. https://doi.org/10.48550/ARXIV.2303.02995.
Gentzkow, Matthew, Jesse M. Shapiro, and Matt Taddy. 2019. “Measuring Group Differences in High‐dimensional Choices: Method and Application to Congressional Speech.” Econometrica 87. https://doi.org/10.3982/ecta16566.
Germain, Mathieu, Karol Gregor, Iain Murray, and Hugo Larochelle. 2015. “MADE: Masked Autoencoder for Distribution Estimation.” arXiv. https://doi.org/10.48550/ARXIV.1502.03509.
Germain, Pascal, Francis Bach, Alexandre Lacoste, and Simon Lacoste-Julien. 2016. “PAC-Bayesian Theory Meets Bayesian Inference.” arXiv. https://doi.org/10.48550/ARXIV.1605.08636.
Geyik, Sahin Cem, Vijay Dialani, Meng Meng, and Ryan Smith. 2018. “In-Session Personalization for Talent Search.” arXiv. https://doi.org/10.48550/ARXIV.1809.06488.
Geyik, Sahin Cem, Sergey Faleev, Jianqiang Shen, Sean O’Donnell, and Santanu Kolay. 2016. “Joint Optimization of Multiple Performance Metrics in Online Video Advertising.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939724.
Geyik, Sahin Cem, Qi Guo, Bo Hu, Cagri Ozcaglar, Ketan Thakkar, Xianren Wu, and Krishnaram Kenthapadi. 2018. “Talent Search and Recommendation Systems at LinkedIn: Practical Challenges and Lessons Learned.” arXiv. https://doi.org/10.48550/ARXIV.1809.06481.
Geyik, Sahin Cem, Abhishek Saxena, and Ali Dasdan. 2014. “Multi-Touch Attribution Based Budget Allocation in Online Advertising.” Proceedings of the Eighth International Workshop on Data Mining for Online Advertising, August. https://doi.org/10.1145/2648584.2648586.
GHAHRAMANI, ZOUBIN. 2001. “AN INTRODUCTION TO HIDDEN MARKOV MODELS AND BAYESIAN NETWORKS.” International Journal of Pattern Recognition and Artificial Intelligence 15 (February). https://doi.org/10.1142/s0218001401000836.
Ghahroodi, Omid, Seyed Arshan Dalili, Sahel Mesforoush, and Ehsaneddin Asgari. 2023. “SUT at SemEval-2023 Task 1: Prompt Generation for Visual Word Sense Disambiguation.” Proceedings of the The 17th International Workshop on Semantic Evaluation (SemEval-2023). https://doi.org/10.18653/v1/2023.semeval-1.298.
Ghasemzadeh, Mohammad, Mohammad Samragh, and Farinaz Koushanfar. 2017. “ReBNet: Residual Binarized Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.1711.01243.
Ghazvininejad, Marjan, Chris Brockett, Ming-Wei Chang, Bill Dolan, Jianfeng Gao, Wen-tau Yih, and Michel Galley. 2017. “A Knowledge-Grounded Neural Conversation Model.” arXiv. https://doi.org/10.48550/ARXIV.1702.01932.
Ghiasi, Golnaz, Honglak Lee, Manjunath Kudlur, Vincent Dumoulin, and Jonathon Shlens. 2017. “Exploring the Structure of a Real-Time, Arbitrary Neural Artistic Stylization Network.” arXiv. https://doi.org/10.48550/ARXIV.1705.06830.
Ghorbani, Amirata, James Wexler, James Zou, and Been Kim. 2019. “Towards Automatic Concept-Based Explanations.” arXiv. https://doi.org/10.48550/ARXIV.1902.03129.
Ghorbani, Amirata, and James Zou. 2019. “Data Shapley: Equitable Valuation of Data for Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1904.02868.
Ghorbani, Saeed, Kimia Mahdaviani, Anne Thaler, Konrad Kording, Douglas James Cook, Gunnar Blohm, and Nikolaus F. Troje. 2021. “MoVi: A Large Multi-Purpose Human Motion and Video Dataset.” PLOS ONE 16 (June). https://doi.org/10.1371/journal.pone.0253157.
Ghosh, Partha, Mehdi S. M. Sajjadi, Antonio Vergari, Michael Black, and Bernhard Schölkopf. 2019. “From Variational to Deterministic Autoencoders.” arXiv. https://doi.org/10.48550/ARXIV.1903.12436.
Ghosh, Riddhi Pratim, Bani Mallick, and Mohsen Pourahmadi. 2021. “Bayesian Estimation of Correlation Matrices of Longitudinal Data.” Bayesian Analysis 16 (September). https://doi.org/10.1214/20-ba1237.
Ghosh, Soumava, Ravi Anand, Tanmoy Bhowmik, and Siddhanth Chandrashekhar. 2023. “GoSage: Heterogeneous Graph Neural Network Using Hierarchical Attention for Collusion Fraud Detection.” 4th ACM International Conference on AI in Finance, November. https://doi.org/10.1145/3604237.3626856.
Ghosh, Tapabrata. 2017. “QuickNet: Maximizing Efficiency and Efficacy in Deep Architectures.” arXiv. https://doi.org/10.48550/ARXIV.1701.02291.
Gidaris, Spyros, and Nikos Komodakis. 2018. “Dynamic Few-Shot Visual Learning Without Forgetting.” arXiv. https://doi.org/10.48550/ARXIV.1804.09458.
———. 2019. “Generating Classification Weights with GNN Denoising Autoencoders for Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.1905.01102.
Gilardi, Fabrizio, Meysam Alizadeh, and Maël Kubli. 2023. “ChatGPT Outperforms Crowd Workers for Text-Annotation Tasks.” Proceedings of the National Academy of Sciences 120 (July). https://doi.org/10.1073/pnas.2305016120.
Gillick, Dan, Cliff Brunk, Oriol Vinyals, and Amarnag Subramanya. 2015. “Multilingual Language Processing from Bytes.” arXiv. https://doi.org/10.48550/ARXIV.1512.00103.
Gilmer, Justin, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. 2017. “Neural Message Passing for Quantum Chemistry.” arXiv. https://doi.org/10.48550/ARXIV.1704.01212.
Gilson, Marion, and Paul Van den Hof. 2005. “Instrumental Variable Methods for Closed-Loop System Identification.” Automatica 41 (February). https://doi.org/10.1016/j.automatica.2004.09.016.
Ginestet, Cedric E., Jun Li, Prakash Balachandran, Steven Rosenberg, and Eric D. Kolaczyk. 2017. “Hypothesis Testing for Network Data in Functional Neuroimaging.” The Annals of Applied Statistics 11 (June). https://doi.org/10.1214/16-aoas1015.
Giordano, Ryan, Runjing Liu, Michael I. Jordan, and Tamara Broderick. 2023. “Evaluating Sensitivity to the Stick-Breaking Prior in Bayesian Nonparametrics (with Discussion).” Bayesian Analysis 18 (March). https://doi.org/10.1214/22-ba1309.
Giorgi, Filippo. 2019. “Thirty Years of Regional Climate Modeling: Where Are We and Where Are We Going Next?” Journal of Geophysical Research: Atmospheres 124 (June). https://doi.org/10.1029/2018jd030094.
Giovannetti, Vittorio, Seth Lloyd, and Lorenzo Maccone. 2004. “Quantum-Enhanced Measurements: Beating the Standard Quantum Limit.” Science 306 (November). https://doi.org/10.1126/science.1104149.
Girish, Deeptha, Vineeta Singh, and Anca Ralescu. 2020. “Understanding Action Recognition in Still Images.” 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), June. https://doi.org/10.1109/cvprw50498.2020.00193.
Giron-Sierra, Jose Maria. 2017. “Digital Signal Processing with Matlab Examples, Volume 3.” Signals and Communication Technology. https://doi.org/10.1007/978-981-10-2540-2.
Girosi, Federico, Michael Jones, and Tomaso Poggio. 1995. “Regularization Theory and Neural Networks Architectures.” Neural Computation 7 (March). https://doi.org/10.1162/neco.1995.7.2.219.
Girshick, Ross. 2015. “Fast r-CNN.” arXiv. https://doi.org/10.48550/ARXIV.1504.08083.
Glenski, Maria, Tim Weninger, and Svitlana Volkova. 2018. “Identifying and Understanding User Reactions to Deceptive and Trusted Social News Sources.” arXiv. https://doi.org/10.48550/ARXIV.1805.12032.
Gliwa, Bogdan, Iwona Mochol, Maciej Biesek, and Aleksander Wawer. 2019. “SAMSum Corpus: A Human-Annotated Dialogue Dataset for Abstractive Summarization.” Proceedings of the 2nd Workshop on New Frontiers in Summarization. https://doi.org/10.18653/v1/d19-5409.
Godichon-Baggioni, Antoine, Nicklas Werge, and Olivier Wintenberger. 2023. “Non-Asymptotic Analysis of Stochastic Approximation Algorithms for Streaming Data.” ESAIM: Probability and Statistics 27. https://doi.org/10.1051/ps/2023006.
Goh, Garrett B., Charles Siegel, Abhinav Vishnu, and Nathan O. Hodas. 2017. “Using Rule-Based Labels for Weak Supervised Learning: A ChemNet for Transferable Chemical Property Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1712.02734.
Goix, Nicolas, Nicolas Drougard, Romain Brault, and Maël Chiapino. 2016. “One Class Splitting Criteria for Random Forests.” arXiv. https://doi.org/10.48550/ARXIV.1611.01971.
Goldberg, David, and James E. Johndrow. 2017. “A Decision Theoretic Approach to a/b Testing.” arXiv. https://doi.org/10.48550/ARXIV.1710.03410.
Goldfeder, Steven, Harry Kalodner, Dillon Reisman, and Arvind Narayanan. 2017. “When the Cookie Meets the Blockchain: Privacy Risks of Web Payments via Cryptocurrencies.” arXiv. https://doi.org/10.48550/ARXIV.1708.04748.
Golding, Andrew R., and Dan Roth. 1998. “A Winnow-Based Approach to Context-Sensitive Spelling Correction.” arXiv. https://doi.org/10.48550/ARXIV.CS/9811003.
Goldman, Claudia V., Martin Allen, and Shlomo Zilberstein. 2006. “Learning to Communicate in a Decentralized Environment.” Autonomous Agents and Multi-Agent Systems 15 (May). https://doi.org/10.1007/s10458-006-0008-9.
Goldman, S. A., and R. H. Sloan. 1995. “Can PAC Learning Algorithms Tolerate Random Attribute Noise?” Algorithmica 14 (July). https://doi.org/10.1007/bf01300374.
Goldstein, Alex, Adam Kapelner, Justin Bleich, and Emil Pitkin. 2013. “Peeking Inside the Black Box: Visualizing Statistical Learning with Plots of Individual Conditional Expectation.” arXiv. https://doi.org/10.48550/ARXIV.1309.6392.
Golub, Maximilian, Guy Lemieux, and Mieszko Lis. 2018. “Full Deep Neural Network Training on a Pruned Weight Budget.” arXiv. https://doi.org/10.48550/ARXIV.1806.06949.
Gomez, Aidan N., Mengye Ren, Raquel Urtasun, and Roger B. Grosse. 2017. “The Reversible Residual Network: Backpropagation Without Storing Activations.” arXiv. https://doi.org/10.48550/ARXIV.1707.04585.
Gómez-Bombarelli, Rafael, Jennifer N. Wei, David Duvenaud, José Miguel Hernández-Lobato, Benjamín Sánchez-Lengeling, Dennis Sheberla, Jorge Aguilera-Iparraguirre, Timothy D. Hirzel, Ryan P. Adams, and Alán Aspuru-Guzik. 2018. “Automatic Chemical Design Using a Data-Driven Continuous Representation of Molecules.” ACS Central Science 4 (January). https://doi.org/10.1021/acscentsci.7b00572.
Gomez-Uribe, Carlos A., and Neil Hunt. 2015. “The Netflix Recommender System.” ACM Transactions on Management Information Systems 6 (December). https://doi.org/10.1145/2843948.
Gong, Shuzhi, Richard O. Sinnott, Jianzhong Qi, and Cecile Paris. 2023. “Fake News Detection Through Graph-Based Neural Networks: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2307.12639.
Gong, Yichen, and Samuel R. Bowman. 2017. “Ruminating Reader: Reasoning with Gated Multi-Hop Attention.” arXiv. https://doi.org/10.48550/ARXIV.1704.07415.
Gong, Yuanhao. 2023. “Dynamic Large Language Models on Blockchains.” arXiv. https://doi.org/10.48550/ARXIV.2307.10549.
Gong, Yunchao, Liu Liu, Ming Yang, and Lubomir Bourdev. 2014. “Compressing Deep Convolutional Networks Using Vector Quantization,” December. http://arxiv.org/abs/1412.6115v1.
Gong, Yu, Yu Zhu, Lu Duan, Qingwen Liu, Ziyu Guan, Fei Sun, Wenwu Ou, and Kenny Q. Zhu. 2019. “Exact-k Recommendation via Maximal Clique Optimization.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330832.
Goodfellow, Ian. 2017. “NIPS 2016 Tutorial: Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1701.00160.
Goodfellow, Ian J., Yaroslav Bulatov, Julian Ibarz, Sacha Arnoud, and Vinay Shet. 2013. “Multi-Digit Number Recognition from Street View Imagery Using Deep Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1312.6082.
Goodfellow, Ian J., Dumitru Erhan, Pierre Luc Carrier, Aaron Courville, Mehdi Mirza, Ben Hamner, Will Cukierski, et al. 2013. “Challenges in Representation Learning: A Report on Three Machine Learning Contests.” arXiv. https://doi.org/10.48550/ARXIV.1307.0414.
Goodfellow, Ian J., Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. “Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1406.2661.
Goodfellow, Ian J., Jonathon Shlens, and Christian Szegedy. 2014. “Explaining and Harnessing Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1412.6572.
Goodfellow, Ian J., David Warde-Farley, Pascal Lamblin, Vincent Dumoulin, Mehdi Mirza, Razvan Pascanu, James Bergstra, Frédéric Bastien, and Yoshua Bengio. 2013. “Pylearn2: A Machine Learning Research Library.” arXiv. https://doi.org/10.48550/ARXIV.1308.4214.
Goodfellow, Ian J., David Warde-Farley, Mehdi Mirza, Aaron Courville, and Yoshua Bengio. 2013. “Maxout Networks.” arXiv. https://doi.org/10.48550/ARXIV.1302.4389.
Goodman, Bryce, and Seth Flaxman. 2017. “European Union Regulations on Algorithmic Decision Making and a ‘Right to Explanation’.” AI Magazine 38 (September). https://doi.org/10.1609/aimag.v38i3.2741.
Goplerud, Max. 2022. “Fast and Accurate Estimation of Non-Nested Binomial Hierarchical Models Using Variational Inference.” Bayesian Analysis 17 (June). https://doi.org/10.1214/21-ba1266.
Gordon, Ariel, Elad Eban, Ofir Nachum, Bo Chen, Hao Wu, Tien-Ju Yang, and Edward Choi. 2017. “MorphNet: Fast &Amp; Simple Resource-Constrained Structure Learning of Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1711.06798.
Gori, M., G. Monfardini, and F. Scarselli. n.d. “A New Model for Learning in Graph Domains.” Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005. https://doi.org/10.1109/ijcnn.2005.1555942.
Gorinova, Maria I., Dave Moore, and Matthew D. Hoffman. 2019. “Automatic Reparameterisation of Probabilistic Programs.” arXiv. https://doi.org/10.48550/ARXIV.1906.03028.
Gormley, Matthew R., Mo Yu, and Mark Dredze. 2015. “Improved Relation Extraction with Feature-Rich Compositional Embedding Models.” arXiv. https://doi.org/10.48550/ARXIV.1505.02419.
Gottardi, Anna, Osman Ipek, Giuseppe Castellucci, Shui Hu, Lavina Vaz, Yao Lu, Anju Khatri, et al. 2022. “Alexa, Let’s Work Together: Introducing the First Alexa Prize TaskBot Challenge on Conversational Task Assistance.” arXiv. https://doi.org/10.48550/ARXIV.2209.06321.
Goyal, Kartik, Graham Neubig, Chris Dyer, and Taylor Berg-Kirkpatrick. 2017. “A Continuous Relaxation of Beam Search for End-to-End Training of Neural Sequence Models.” arXiv. https://doi.org/10.48550/ARXIV.1708.00111.
Goyal, Palash, and Emilio Ferrara. 2018. “Graph Embedding Techniques, Applications, and Performance: A Survey.” Knowledge-Based Systems 151 (July). https://doi.org/10.1016/j.knosys.2018.03.022.
Goyal, Priya, Piotr Dollár, Ross Girshick, Pieter Noordhuis, Lukasz Wesolowski, Aapo Kyrola, Andrew Tulloch, Yangqing Jia, and Kaiming He. 2017. “Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour.” arXiv. https://doi.org/10.48550/ARXIV.1706.02677.
Goyal, Yash, Ziyan Wu, Jan Ernst, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019. “Counterfactual Visual Explanations.” arXiv. https://doi.org/10.48550/ARXIV.1904.07451.
Gozalo-Brizuela, Roberto, and Eduardo C. Garrido-Merchan. 2023. “ChatGPT Is Not All You Need. A State of the Art Review of Large Generative AI Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.04655.
Grachev, Artem M., Dmitry I. Ignatov, and Andrey V. Savchenko. 2019. “Compression of Recurrent Neural Networks for Efficient Language Modeling.” Applied Soft Computing 79 (June). https://doi.org/10.1016/j.asoc.2019.03.057.
Graham, Benjamin. 2014. “Spatially-Sparse Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1409.6070.
———. 2017. “Low-Precision Batch-Normalized Activations.” arXiv. https://doi.org/10.48550/ARXIV.1702.08231.
Graham, Simon, Fayyaz Minhas, Mohsin Bilal, Mahmoud Ali, Yee Wah Tsang, Mark Eastwood, Noorul Wahab, et al. 2023. “Screening of Normal Endoscopic Large Bowel Biopsies with Interpretable Graph Learning: A Retrospective Study.” Gut 72 (May). https://doi.org/10.1136/gutjnl-2023-329512.
Graikos, Alexandros, Nikolay Malkin, Nebojsa Jojic, and Dimitris Samaras. 2022. “Diffusion Models as Plug-and-Play Priors.” arXiv. https://doi.org/10.48550/ARXIV.2206.09012.
Gramacy, Robert B., and Herbert K. H. Lee. 2007. “Bayesian Treed Gaussian Process Models with an Application to Computer Modeling.” arXiv. https://doi.org/10.48550/ARXIV.0710.4536.
Graves, Alex. 2012. “Supervised Sequence Labelling with Recurrent Neural Networks.” Studies in Computational Intelligence. https://doi.org/10.1007/978-3-642-24797-2.
———. 2013. “Generating Sequences with Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1308.0850.
———. 2016. “Adaptive Computation Time for Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1603.08983.
Graves, Alex, Marc G. Bellemare, Jacob Menick, Remi Munos, and Koray Kavukcuoglu. 2017. “Automated Curriculum Learning for Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1704.03003.
Graves, Alex, Abdel-rahman Mohamed, and Geoffrey Hinton. 2013. “Speech Recognition with Deep Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1303.5778.
Graves, Alex, Greg Wayne, and Ivo Danihelka. 2014. “Neural Turing Machines.” arXiv. https://doi.org/10.48550/ARXIV.1410.5401.
Graves, Alex, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwińska, Sergio Gómez Colmenarejo, et al. 2016. “Hybrid Computing Using a Neural Network with Dynamic External Memory.” Nature 538 (October). https://doi.org/10.1038/nature20101.
Gray, Robert M. 2005. “Toeplitz and Circulant Matrices: A Review.” Foundations and Trends® in Communications and Information Theory 2. https://doi.org/10.1561/0100000006.
Grbovic, Mihajlo, Vladan Radosavljevic, Nemanja Djuric, Narayan Bhamidipati, Jaikit Savla, Varun Bhagwan, and Doug Sharp. 2015. “E-Commerce in Your Inbox.” Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2783258.2788627.
Greene, Derek, and James P. Cross. 2016. “Exploring the Political Agenda of the European Parliament Using a Dynamic Topic Modeling Approach.” arXiv. https://doi.org/10.48550/ARXIV.1607.03055.
Gregor, Karol, Ivo Danihelka, Alex Graves, Danilo Jimenez Rezende, and Daan Wierstra. 2015. “DRAW: A Recurrent Neural Network for Image Generation.” arXiv. https://doi.org/10.48550/ARXIV.1502.04623.
Gregor, Karo, and Yann LeCun. 2010. “Emergence of Complex-Like Cells in a Temporal Product Network with Local Receptive Fields.” arXiv. https://doi.org/10.48550/ARXIV.1006.0448.
Gretton, Arthur, Karsten Borgwardt, Malte J. Rasch, Bernhard Scholkopf, and Alexander J. Smola. 2008. “A Kernel Method for the Two-Sample Problem.” arXiv. https://doi.org/10.48550/ARXIV.0805.2368.
Greven, Andreas, Peter Pfaffelhuber, and Anita Winter. 2008. “Convergence in Distribution of Random Metric Measure Spaces (λ-Coalescent Measure Trees).” Probability Theory and Related Fields 145 (August). https://doi.org/10.1007/s00440-008-0169-3.
Grigas, Paul, Alfonso Lobos, Zheng Wen, and Kuang-chih Lee. 2017. “Profit Maximization for Online Advertising Demand-Side Platforms.” arXiv. https://doi.org/10.48550/ARXIV.1706.01614.
Grigg, Tom George, Dan Busbridge, Jason Ramapuram, and Russ Webb. 2021. “Do Self-Supervised and Supervised Methods Learn Similar Visual Representations?” arXiv. https://doi.org/10.48550/ARXIV.2110.00528.
Grill, Jean-Bastien, Florent Altché, Yunhao Tang, Thomas Hubert, Michal Valko, Ioannis Antonoglou, and Rémi Munos. 2020. “Monte-Carlo Tree Search as Regularized Policy Optimization.” arXiv. https://doi.org/10.48550/ARXIV.2007.12509.
Grimmer, Justin, and Gary King. 2011. “General Purpose Computer-Assisted Clustering and Conceptualization.” Proceedings of the National Academy of Sciences 108 (February). https://doi.org/10.1073/pnas.1018067108.
Grinsztajn, Léo, Elizaveta Semenova, Charles C. Margossian, and Julien Riou. 2020. “Bayesian Workflow for Disease Transmission Modeling in Stan.” arXiv. https://doi.org/10.48550/ARXIV.2006.02985.
Groeneboom, Piet. 1989. “Brownian Motion with a Parabolic Drift and Airy Functions.” Probability Theory and Related Fields 81 (February). https://doi.org/10.1007/bf00343738.
Gromenko, Oleksandr, Piotr Kokoszka, Lie Zhu, and Jan Sojka. 2012. “Estimation and Testing for Spatially Indexed Curves with Application to Ionospheric and Magnetic Field Trends.” The Annals of Applied Statistics 6 (June). https://doi.org/10.1214/11-aoas524.
Grosse, Roger, Juhan Bae, Cem Anil, Nelson Elhage, Alex Tamkin, Amirhossein Tajdini, Benoit Steiner, et al. 2023. “Studying Large Language Model Generalization with Influence Functions.” arXiv. https://doi.org/10.48550/ARXIV.2308.03296.
Grover, Aditya, and Stefano Ermon. 2017. “Boosted Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.1702.08484.
Grover, Aditya, and Jure Leskovec. 2016. “Node2vec.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939754.
Grubb, Alexander, and J. Andrew Bagnell. 2011. “Generalized Boosting Algorithms for Convex Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1105.2054.
Gruslys, Audrunas, Will Dabney, Mohammad Gheshlaghi Azar, Bilal Piot, Marc Bellemare, and Remi Munos. 2017. “The Reactor: A Fast and Sample-Efficient Actor-Critic Agent for Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1704.04651.
Grzech, Daniel, Mohammad Farid Azampour, Huaqi Qiu, Ben Glocker, Bernhard Kainz, and Loïc Le Folgoc. 2021. “Uncertainty Quantification in Non-Rigid Image Registration via Stochastic Gradient Markov Chain Monte Carlo.” arXiv. https://doi.org/10.48550/ARXIV.2110.13289.
Gu, Jindong, Zhen Han, Shuo Chen, Ahmad Beirami, Bailan He, Gengyuan Zhang, Ruotong Liao, Yao Qin, Volker Tresp, and Philip Torr. 2023. “A Systematic Survey of Prompt Engineering on Vision-Language Foundation Models.” arXiv. https://doi.org/10.48550/ARXIV.2307.12980.
Gu, Mengyang, and Hanmo Li. 2022. “Gaussian Orthogonal Latent Factor Processes for Large Incomplete Matrices of Correlated Data.” Bayesian Analysis 17 (December). https://doi.org/10.1214/21-ba1295.
Gu, Shixiang, Timothy Lillicrap, Ilya Sutskever, and Sergey Levine. 2016. “Continuous Deep q-Learning with Model-Based Acceleration.” arXiv. https://doi.org/10.48550/ARXIV.1603.00748.
Gu, Shixiang, and Luca Rigazio. 2014. “Towards Deep Neural Network Architectures Robust to Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1412.5068.
Gu, Xiaodong, Hongyu Zhang, Dongmei Zhang, and Sunghun Kim. 2016. “Deep API Learning.” arXiv. https://doi.org/10.48550/ARXIV.1605.08535.
———. 2017. “DeepAM: Migrate APIs with Multi-Modal Sequence to Sequence Learning.” arXiv. https://doi.org/10.48550/ARXIV.1704.07734.
Gu, Yupeng, Yizhou Sun, Yanen Li, and Yang Yang. 2018. “RaRE.” Proceedings of the 2018 World Wide Web Conference on World Wide Web - WWW ’18. https://doi.org/10.1145/3178876.3186102.
Gu, Yuxian, Jiaxin Wen, Hao Sun, Yi Song, Pei Ke, Chujie Zheng, Zheng Zhang, et al. 2023. “EVA2.0: Investigating Open-Domain Chinese Dialogue Systems with Large-Scale Pre-Training.” Machine Intelligence Research 20 (February). https://doi.org/10.1007/s11633-022-1387-3.
Guan, Melody Y., Varun Gulshan, Andrew M. Dai, and Geoffrey E. Hinton. 2017. “Who Said What: Modeling Individual Labelers Improves Classification.” arXiv. https://doi.org/10.48550/ARXIV.1703.08774.
Guclu, U., and M. A. J. van Gerven. 2015. “Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations Across the Ventral Stream.” Journal of Neuroscience 35 (July). https://doi.org/10.1523/jneurosci.5023-14.2015.
Guerrero, Paul, Yanir Kleiman, Maks Ovsjanikov, and Niloy J. Mitra. 2018. “PCPN<scp>et</Scp> Learning Local Shape Properties from Raw Point Clouds.” Computer Graphics Forum 37 (May). https://doi.org/10.1111/cgf.13343.
Guha, Neel, Julian Nyarko, Daniel E. Ho, Christopher Ré, Adam Chilton, Aditya Narayana, Alex Chohlas-Wood, et al. 2023. “LegalBench: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2308.11462.
Guhaniyogi, Rajarshi, and Daniel Spencer. 2021. “Bayesian Tensor Response Regression with an Application to Brain Activation Studies.” Bayesian Analysis 16 (December). https://doi.org/10.1214/21-ba1280.
Gulcehre, Caglar, Orhan Firat, Kelvin Xu, Kyunghyun Cho, Loic Barrault, Huei-Chi Lin, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2015. “On Using Monolingual Corpora in Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1503.03535.
Gunawan, Alexander Agung Santoso, Ananda Phan Iman, and Derwin Suhartono. 2020. “Automatic Music Generator Using Recurrent Neural Network.” International Journal of Computational Intelligence Systems 13. https://doi.org/10.2991/ijcis.d.200519.001.
Gunluk, Oktay, Jayant Kalagnanam, Minhan Li, Matt Menickelly, and Katya Scheinberg. 2016. “Optimal Generalized Decision Trees via Integer Programming.” arXiv. https://doi.org/10.48550/ARXIV.1612.03225.
Guo, Chuan, Mayank Rana, Moustapha Cisse, and Laurens van der Maaten. 2017. “Countering Adversarial Images Using Input Transformations.” arXiv. https://doi.org/10.48550/ARXIV.1711.00117.
Guo, Haohan, Fenglong Xie, Frank K. Soong, Xixin Wu, and Helen Meng. 2022. “A Multi-Stage Multi-Codebook VQ-VAE Approach to High-Performance Neural TTS.” arXiv. https://doi.org/10.48550/ARXIV.2209.10887.
Guo, Jian, and Stephen Gould. 2015. “Deep CNN Ensemble with Data Augmentation for Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.1506.07224.
Guo, Jiang, Darsh J Shah, and Regina Barzilay. 2018. “Multi-Source Domain Adaptation with Mixture of Experts.” arXiv. https://doi.org/10.48550/ARXIV.1809.02256.
Guo, Junliang, Linli Xu, and Jingchang Liu. 2018. “SPINE: Structural Identity Preserved Inductive Network Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1802.03984.
Guo, Lin Lawrence, Ethan Steinberg, Scott Lanyon Fleming, Jose Posada, Joshua Lemmon, Stephen R. Pfohl, Nigam Shah, Jason Fries, and Lillian Sung. 2023. “EHR Foundation Models Improve Robustness in the Presence of Temporal Distribution Shift.” Scientific Reports 13 (March). https://doi.org/10.1038/s41598-023-30820-8.
Guo, Pei, and Ryan Farrell. 2018. “Semantic Network Interpretation.” arXiv. https://doi.org/10.48550/ARXIV.1805.08969.
Guo, Qingyan, Rui Wang, Junliang Guo, Bei Li, Kaitao Song, Xu Tan, Guoqing Liu, Jiang Bian, and Yujiu Yang. 2023. “Connecting Large Language Models with Evolutionary Algorithms Yields Powerful Prompt Optimizers.” arXiv. https://doi.org/10.48550/ARXIV.2309.08532.
Guo, Ruiqi, Philip Sun, Erik Lindgren, Quan Geng, David Simcha, Felix Chern, and Sanjiv Kumar. 2019. “Accelerating Large-Scale Inference with Anisotropic Vector Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1908.10396.
Guo, Shu, Quan Wang, Lihong Wang, Bin Wang, and Li Guo. 2017. “Knowledge Graph Embedding with Iterative Guidance from Soft Rules.” arXiv. https://doi.org/10.48550/ARXIV.1711.11231.
Guo, Xingzhi, Baojian Zhou, and Steven Skiena. 2022. “Subset Node Anomaly Tracking over Large Dynamic Graphs.” Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3534678.3539389.
Guo, Yiwen, Anbang Yao, and Yurong Chen. 2016. “Dynamic Network Surgery for Efficient DNNs.” arXiv. https://doi.org/10.48550/ARXIV.1608.04493.
Guo, Yunhui, Haoran Guo, and Stella Yu. 2021. “CO-SNE: Dimensionality Reduction and Visualization for Hyperbolic Data.” arXiv. https://doi.org/10.48550/ARXIV.2111.15037.
Gupta, Agrim, Justin Johnson, Alexandre Alahi, and Li Fei-Fei. 2017. “Characterizing and Improving Stability in Neural Style Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1705.02092.
Gupta, Manish, Jing Gao, Charu C. Aggarwal, and Jiawei Han. 2014. “Outlier Detection for Temporal Data: A Survey.” IEEE Transactions on Knowledge and Data Engineering 26 (September). https://doi.org/10.1109/tkde.2013.184.
Gupta, Saurabh, Judy Hoffman, and Jitendra Malik. 2015. “Cross Modal Distillation for Supervision Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1507.00448.
Gupta, Sonal, Rushin Shah, Mrinal Mohit, Anuj Kumar, and Mike Lewis. 2018. “Semantic Parsing for Task Oriented Dialog Using Hierarchical Representations.” arXiv. https://doi.org/10.48550/ARXIV.1810.07942.
Gupta, Suyog, Ankur Agrawal, Kailash Gopalakrishnan, and Pritish Narayanan. 2015. “Deep Learning with Limited Numerical Precision.” arXiv. https://doi.org/10.48550/ARXIV.1502.02551.
Gupta, Udit, Brandon Reagen, Lillian Pentecost, Marco Donato, Thierry Tambe, Alexander M. Rush, Gu-Yeon Wei, and David Brooks. 2019. “MASR: A Modular Accelerator for Sparse RNNs.” arXiv. https://doi.org/10.48550/ARXIV.1908.08976.
Gurumoorthy, Karthik S., Subhajit Sanyal, and Vineet Chaoji. 2020. “Think Out of the Package: Recommending Package Types for e-Commerce Shipments.” arXiv. https://doi.org/10.48550/ARXIV.2006.03239.
Gutman, David, Noel C. F. Codella, Emre Celebi, Brian Helba, Michael Marchetti, Nabin Mishra, and Allan Halpern. 2016. “Skin Lesion Analysis Toward Melanoma Detection: A Challenge at the International Symposium on Biomedical Imaging (ISBI) 2016, Hosted by the International Skin Imaging Collaboration (ISIC).” arXiv. https://doi.org/10.48550/ARXIV.1605.01397.
Guu, Kelvin, John Miller, and Percy Liang. 2015. “Traversing Knowledge Graphs in Vector Space.” arXiv. https://doi.org/10.48550/ARXIV.1506.01094.
Guzzi, P. H., M. Mina, C. Guerra, and M. Cannataro. 2011. “Semantic Similarity Analysis of Protein Data: Assessment with Biological Features and Issues.” Briefings in Bioinformatics 13 (December). https://doi.org/10.1093/bib/bbr066.
Gwee, Xian Yao, Isobel Claire Gormley, and Michael Fop. 2022. “A Latent Shrinkage Position Model for Binary and Count Network Data.” arXiv. https://doi.org/10.48550/ARXIV.2211.13034.
Gwilliams, Laura, Graham Flick, Alec Marantz, Liina Pylkkänen, David Poeppel, and Jean-Rémi King. 2023. “Introducing MEG-MASC a High-Quality Magneto-Encephalography Dataset for Evaluating Natural Speech Processing.” Scientific Data 10 (December). https://doi.org/10.1038/s41597-023-02752-5.
Gyöngy, I., and N. V. Krylov. 2021. “Existence of Strong Solutions for Itô’s Stochastic Equations via Approximations. Revisited.” arXiv. https://doi.org/10.48550/ARXIV.2107.14384.
Gysel, Philipp, Mohammad Motamedi, and Soheil Ghiasi. 2016. “Hardware-Oriented Approximation of Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1604.03168.
Ha, David, Andrew Dai, and Quoc V. Le. 2016. “HyperNetworks.” arXiv. https://doi.org/10.48550/ARXIV.1609.09106.
Ha, David, and Douglas Eck. 2017. “A Neural Representation of Sketch Drawings.” arXiv. https://doi.org/10.48550/ARXIV.1704.03477.
Ha, Jung-Woo, Hyuna Pyo, and Jeonghee Kim. 2016. “Large-Scale Item Categorization in e-Commerce Using Multiple Recurrent Neural Networks.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939678.
Haas, Moritz, David Holzmüller, Ulrike von Luxburg, and Ingo Steinwart. 2023. “Mind the Spikes: Benign Overfitting of Kernels and Neural Networks in Fixed Dimension.” arXiv. https://doi.org/10.48550/ARXIV.2305.14077.
Hadfield-Menell, Dylan, Anca Dragan, Pieter Abbeel, and Stuart Russell. 2016. “Cooperative Inverse Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.03137.
Hadi, Mohammad Abdul, and Fatemeh H. Fard. 2021. “Evaluating Pre-Trained Models for User Feedback Analysis in Software Engineering: A Study on Classification of App-Reviews.” arXiv. https://doi.org/10.48550/ARXIV.2104.05861.
Haffari, Gholamreza, Yang Wang, Shaojun Wang, Greg Mori, and Feng Jiao. 2008. “Boosting with Incomplete Information.” Proceedings of the 25th International Conference on Machine Learning - ICML ’08. https://doi.org/10.1145/1390156.1390203.
Hainmueller, Jens, and Chad Hazlett. 2014. “Kernel Regularized Least Squares: Reducing Misspecification Bias with a Flexible and Interpretable Machine Learning Approach.” Political Analysis 22. https://doi.org/10.1093/pan/mpt019.
Hajič, Jan, and Pavel Pecina. 2017. “In Search of a Dataset for Handwritten Optical Music Recognition: Introducing MUSCIMA++.” arXiv. https://doi.org/10.48550/ARXIV.1703.04824.
Hakkak, Hamed. 2018. “Auto Deep Compression by Reinforcement Learning Based Actor-Critic Structure.” arXiv. https://doi.org/10.48550/ARXIV.1807.02886.
Haldar, Malay, Mustafa Abdool, Prashant Ramanathan, Tyler Sax, Lanbo Zhang, Aamir Mansawala, Shulin Yang, Bradley Turnbull, and Junshuo Liao. 2020. “Improving Deep Learning for Airbnb Search.” arXiv. https://doi.org/10.48550/ARXIV.2002.05515.
Haldar, Malay, Mustafa Abdool, Prashant Ramanathan, Tao Xu, Shulin Yang, Huizhong Duan, Qing Zhang, et al. 2019. “Applying Deep Learning to Airbnb Search.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330658.
Hall, Peter, and James Stephen Marron. 1987. “Extent to Which Least-Squares Cross-Validation Minimises Integrated Square Error in Nonparametric Density Estimation.” Probability Theory and Related Fields 74 (April). https://doi.org/10.1007/bf00363516.
Hall, Peter, and Prakash Patil. 1994. “Properties of Nonparametric Estimators of Autocovariance for Stationary Random Fields.” Probability Theory and Related Fields 99 (September). https://doi.org/10.1007/bf01199899.
Hämäläinen, Perttu, Mikke Tavast, and Anton Kunnari. 2023. “Evaluating Large Language Models in Generating Synthetic HCI Research Data: A Case Study.” Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, April. https://doi.org/10.1145/3544548.3580688.
Hamilton, William L., Kevin Clark, Jure Leskovec, and Dan Jurafsky. 2016. “Inducing Domain-Specific Sentiment Lexicons from Unlabeled Corpora.” arXiv. https://doi.org/10.48550/ARXIV.1606.02820.
Hamilton, William L., Rex Ying, and Jure Leskovec. 2017a. “Inductive Representation Learning on Large Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1706.02216.
———. 2017b. “Representation Learning on Graphs: Methods and Applications.” arXiv. https://doi.org/10.48550/ARXIV.1709.05584.
Hamooni, Hossein, Biplob Debnath, Jianwu Xu, Hui Zhang, Guofei Jiang, and Abdullah Mueen. 2016. “LogMine.” Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, October. https://doi.org/10.1145/2983323.2983358.
Hamrick, Jessica B., Kelsey R. Allen, Victor Bapst, Tina Zhu, Kevin R. McKee, Joshua B. Tenenbaum, and Peter W. Battaglia. 2018. “Relational Inductive Bias for Physical Construction in Humans and Machines.” arXiv. https://doi.org/10.48550/ARXIV.1806.01203.
Han, Chi, Qifan Wang, Wenhan Xiong, Yu Chen, Heng Ji, and Sinong Wang. 2023. “LM-Infinite: Simple on-the-Fly Length Generalization for Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2308.16137.
Han, Ligong, Yinxiao Li, Han Zhang, Peyman Milanfar, Dimitris Metaxas, and Feng Yang. 2023. “SVDiff: Compact Parameter Space for Diffusion Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2303.11305.
Han, Rui, and Fan Yang. 2024. “Decay of Multi-Point Correlation Functions in \[\mathbb {Z}^d\].” Communications in Mathematical Physics 405 (February). https://doi.org/10.1007/s00220-023-04884-6.
Han, Seungyeop, Haichen Shen, Matthai Philipose, Sharad Agarwal, Alec Wolman, and Arvind Krishnamurthy. 2016. “MCDNN.” Proceedings of the 14th Annual International Conference on Mobile Systems, Applications, and Services, June. https://doi.org/10.1145/2906388.2906396.
Han, Song, Xingyu Liu, Huizi Mao, Jing Pu, Ardavan Pedram, Mark A. Horowitz, and William J. Dally. 2016. “EIE: Efficient Inference Engine on Compressed Deep Neural Network,” February. http://arxiv.org/abs/1602.01528v2.
Han, Song, Huizi Mao, and William J. Dally. 2015. “Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding.” arXiv. https://doi.org/10.48550/ARXIV.1510.00149.
Han, Song, Jeff Pool, Sharan Narang, Huizi Mao, Enhao Gong, Shijian Tang, Erich Elsen, et al. 2016. “DSD: Dense-Sparse-Dense Training for Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1607.04381.
Han, Song, Jeff Pool, John Tran, and William J. Dally. 2015. “Learning Both Weights and Connections for Efficient Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1506.02626.
Han, Xiao, Xiatian Zhu, Licheng Yu, Li Zhang, Yi-Zhe Song, and Tao Xiang. 2023. “FAME-ViL: Multi-Tasking Vision-Language Model for Heterogeneous Fashion Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2303.02483.
Hand, David J., and Keming Yu. 2001. “Idiot’s Bayes: Not so Stupid After All?” International Statistical Review / Revue Internationale de Statistique 69 (December). https://doi.org/10.2307/1403452.
“Handbook of Digital Face Manipulation and Detection.” 2022. Advances in Computer Vision and Pattern Recognition. https://doi.org/10.1007/978-3-030-87664-7.
Hannun, Awni Y., Andrew L. Maas, Daniel Jurafsky, and Andrew Y. Ng. 2014. “First-Pass Large Vocabulary Continuous Speech Recognition Using Bi-Directional Recurrent DNNs.” arXiv. https://doi.org/10.48550/ARXIV.1408.2873.
Hannun, Awni, Carl Case, Jared Casper, Bryan Catanzaro, Greg Diamos, Erich Elsen, Ryan Prenger, et al. 2014. “Deep Speech: Scaling up End-to-End Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1412.5567.
Hansen, Stine, Srishti Gautam, Suaiba Amina Salahuddin, Michael Kampffmeyer, and Robert Jenssen. 2023. “ADNet++: A Few-Shot Learning Framework for Multi-Class Medical Image Volume Segmentation with Uncertainty-Guided Feature Refinement.” Medical Image Analysis 89 (October). https://doi.org/10.1016/j.media.2023.102870.
Hao, Jianye, Dongping Huang, Yi Cai, and Ho-Fung Leung. 2014. “Networked Reinforcement Social Learning Towards Coordination in Cooperative Multiagent Systems.” 2014 IEEE 26th International Conference on Tools with Artificial Intelligence, November. https://doi.org/10.1109/ictai.2014.63.
Hao, Junheng, Tong Zhao, Jin Li, Xin Luna Dong, Christos Faloutsos, Yizhou Sun, and Wei Wang. 2020. “P-Companion.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412732.
Hao, Tianxiang, Mengyao Lyu, Hui Chen, Sicheng Zhao, Jungong Han, and Guiguang Ding. 2023. “Re-Parameterized Low-Rank Prompt: Generalize a Vision-Language Model Within 0.5K Parameters.” arXiv. https://doi.org/10.48550/ARXIV.2312.10813.
Haonan, Lu, Seth H. Huang, Tian Ye, and Guo Xiuyan. 2019. “Graph Star Net for Generalized Multi-Task Learning.” arXiv. https://doi.org/10.48550/ARXIV.1906.12330.
Haotian, Hu, Wang Fanyi, Su Jingwen, Gao Shiyu, and Zhang Zhiwang. 2023. “IC-FPS: Instance-Centroid Faster Point Sampling Module for 3D Point-Base Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2303.17921.
Hardin, Johanna, Stephan Ramon Garcia, and David Golan. 2013. “A Method for Generating Realistic Correlation Matrices.” The Annals of Applied Statistics 7 (September). https://doi.org/10.1214/13-aoas638.
Hariharan, Bharath, and Ross Girshick. 2016. “Low-Shot Visual Recognition by Shrinking and Hallucinating Features.” arXiv. https://doi.org/10.48550/ARXIV.1606.02819.
Hariri, Walid. 2023. “Unlocking the Potential of ChatGPT: A Comprehensive Exploration of Its Applications, Advantages, Limitations, and Future Directions in Natural Language Processing.” arXiv. https://doi.org/10.48550/ARXIV.2304.02017.
Harrow, Aram W., Avinatan Hassidim, and Seth Lloyd. 2009. “Quantum Algorithm for Linear Systems of Equations.” Physical Review Letters 103 (October). https://doi.org/10.1103/physrevlett.103.150502.
Hartmann, Jochen, Jasper Schwenzow, and Maximilian Witte. 2023. “The Political Ideology of Conversational AI: Converging Evidence on ChatGPT’s Pro-Environmental, Left-Libertarian Orientation.” arXiv. https://doi.org/10.48550/ARXIV.2301.01768.
Hashemi, Soheil, Nicholas Anthony, Hokchhay Tann, R. Iris Bahar, and Sherief Reda. 2016. “Understanding the Impact of Precision Quantization on the Accuracy and Energy of Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1612.03940.
Haskell, Karen H., and Richard J. Hanson. 1981. “An Algorithm for Linear Least Squares Problems with Equality and Nonnegativity Constraints.” Mathematical Programming 21 (December). https://doi.org/10.1007/bf01584232.
Haslbeck, Jonas M. B., and Lourens J. Waldorp. 2017. “How Well Do Network Models Predict Observations? On the Importance of Predictability in Network Models.” Behavior Research Methods 50 (July). https://doi.org/10.3758/s13428-017-0910-x.
Hassani, Hossein, and Emmanuel Sirmal Silva. 2023. “The Role of ChatGPT in Data Science: How AI-Assisted Conversational Interfaces Are Revolutionizing the Field.” Big Data and Cognitive Computing 7 (March). https://doi.org/10.3390/bdcc7020062.
Hasselt, Hado van, Arthur Guez, and David Silver. 2015. “Deep Reinforcement Learning with Double q-Learning.” arXiv. https://doi.org/10.48550/ARXIV.1509.06461.
Hastie, Trevor, Rahul Mazumder, Jason Lee, and Reza Zadeh. 2014. “Matrix Completion and Low-Rank SVD via Fast Alternating Least Squares.” arXiv. https://doi.org/10.48550/ARXIV.1410.2596.
Hastie, Trevor, Andrea Montanari, Saharon Rosset, and Ryan J. Tibshirani. 2019. “Surprises in High-Dimensional Ridgeless Least Squares Interpolation.” arXiv. https://doi.org/10.48550/ARXIV.1903.08560.
Hastie, Trevor, and Werner Stuetzle. 1989. “Principal Curves.” Journal of the American Statistical Association 84 (June). https://doi.org/10.1080/01621459.1989.10478797.
Hata, Hideaki, Emad Shihab, and Graham Neubig. 2018. “Learning to Generate Corrective Patches Using Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1812.07170.
Ha-Thuc, Viet, and Shakti Sinha. 2016. “Learning to Rank Personalized Search Results in Professional Networks.” Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/2911451.2927018.
Haug, Charlotte J., and Jeffrey M. Drazen. 2023. “Artificial Intelligence and Machine Learning in Clinical Medicine, 2023.” New England Journal of Medicine 388 (March). https://doi.org/10.1056/nejmra2302038.
Havasi, Marton, Robert Peharz, and José Miguel Hernández-Lobato. 2018. “Minimal Random Code Learning: Getting Bits Back from Compressed Model Parameters.” arXiv. https://doi.org/10.48550/ARXIV.1810.00440.
Hawthorne, Curtis, Ian Simon, Rigel Swavely, Ethan Manilow, and Jesse Engel. 2021. “Sequence-to-Sequence Piano Transcription with Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2107.09142.
Hayashi, Hiroaki, Jayanth Koushik, and Graham Neubig. 2016. “Eve: A Gradient Based Optimization Method with Locally and Globally Adaptive Learning Rates.” arXiv. https://doi.org/10.48550/ARXIV.1611.01505.
Hayashi-Takagi, Akiko, Sho Yagishita, Mayumi Nakamura, Fukutoshi Shirai, Yi I. Wu, Amanda L. Loshbaugh, Brian Kuhlman, Klaus M. Hahn, and Haruo Kasai. 2015. “Labelling and Optical Erasure of Synaptic Memory Traces in the Motor Cortex.” Nature 525 (September). https://doi.org/10.1038/nature15257.
Hays, Spencer, Haipeng Shen, and Jianhua Z. Huang. 2012. “Functional Dynamic Factor Models with Application to Yield Curve Forecasting.” The Annals of Applied Statistics 6 (September). https://doi.org/10.1214/12-aoas551.
Hazan, Elad, and Satyen Kale. 2010. “An Optimal Algorithm for Stochastic Strongly-Convex Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1006.2425.
———. 2012. “Projection-Free Online Learning.” arXiv. https://doi.org/10.48550/ARXIV.1206.4657.
Hazra, Arnab, and Raphaël Huser. 2021. “Estimating High-Resolution Red Sea Surface Temperature Hotspots, Using a Low-Rank Semiparametric Spatial Model.” The Annals of Applied Statistics 15 (June). https://doi.org/10.1214/20-aoas1418.
He, Congqing, Li Peng, Yuquan Le, Jiawei He, and Xiangyu Zhu. 2019. “SECaps: A Sequence Enhanced Capsule Model for Charge Prediction.” Artificial Neural Networks and Machine Learning – ICANN 2019: Text and Time Series. https://doi.org/10.1007/978-3-030-30490-4_19.
He, Hengtao, Shi Jin, Chao-Kai Wen, Feifei Gao, Geoffrey Ye Li, and Zongben Xu. 2018. “Model-Driven Deep Learning for Physical Layer Communications.” arXiv. https://doi.org/10.48550/ARXIV.1809.06059.
He, Hua, Kevin Gimpel, and Jimmy Lin. 2015. “Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks.” Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/d15-1181.
He, Jingxuan, Mislav Balunović, Nodar Ambroladze, Petar Tsankov, and Martin Vechev. 2019. “Learning to Fuzz from Symbolic Execution with Application to Smart Contracts.” Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, November. https://doi.org/10.1145/3319535.3363230.
He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2015a. “Deep Residual Learning for Image Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1512.03385.
———. 2015b. “Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification.” arXiv. https://doi.org/10.48550/ARXIV.1502.01852.
———. 2016. “Identity Mappings in Deep Residual Networks.” arXiv. https://doi.org/10.48550/ARXIV.1603.05027.
He, Qiang, Guowei Chen, Wenchao Song, and Pengzhou Zhang. 2023. “Prompt-Based Word-Level Information Injection BERT for Chinese Named Entity Recognition.” Applied Sciences 13 (March). https://doi.org/10.3390/app13053331.
He, Qinyao, He Wen, Shuchang Zhou, Yuxin Wu, Cong Yao, Xinyu Zhou, and Yuheng Zou. 2016. “Effective Quantization Methods for Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.10176.
He, Ruining, Wang-Cheng Kang, and Julian McAuley. 2017. “Translation-Based Recommendation.” Proceedings of the Eleventh ACM Conference on Recommender Systems, August. https://doi.org/10.1145/3109859.3109882.
He, Tao, Lianli Gao, Jingkuan Song, and Yuan-Fang Li. 2022. “Towards Open-Vocabulary Scene Graph Generation with Prompt-Based Finetuning.” arXiv. https://doi.org/10.48550/ARXIV.2208.08165.
He, Tong, Weilin Huang, Yu Qiao, and Jian Yao. 2016. “Accurate Text Localization in Natural Image with Cascaded Convolutional Text Network.” arXiv. https://doi.org/10.48550/ARXIV.1603.09423.
He, Tong, Zhi Zhang, Hang Zhang, Zhongyue Zhang, Junyuan Xie, and Mu Li. 2018. “Bag of Tricks for Image Classification with Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1812.01187.
He, Warren, James Wei, Xinyun Chen, Nicholas Carlini, and Dawn Song. 2017. “Adversarial Example Defenses: Ensembles of Weak Defenses Are Not Strong.” arXiv. https://doi.org/10.48550/ARXIV.1706.04701.
He, Xiangheng, Junjie Chen, Georgios Rizos, and Björn W. Schuller. 2021. “An Improved StarGAN for Emotional Voice Conversion: Enhancing Voice Quality and Data Augmentation.” arXiv. https://doi.org/10.48550/ARXIV.2107.08361.
He, Xiangnan, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. 2017. “Neural Collaborative Filtering.” arXiv. https://doi.org/10.48550/ARXIV.1708.05031.
He, Xiangnan, Yang Zhang, Fuli Feng, Chonggang Song, Lingling Yi, Guohui Ling, and Yongdong Zhang. 2022. “Addressing Confounding Feature Issue for Causal Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.2205.06532.
He, Xiaoxi, Dawei Gao, Zimu Zhou, Yongxin Tong, and Lothar Thiele. 2019. “Pruning-Aware Merging for Efficient Multitask Inference.” arXiv. https://doi.org/10.48550/ARXIV.1905.09676.
He, Xinlei, Xinyue Shen, Zeyuan Chen, Michael Backes, and Yang Zhang. 2023. “MGTBench: Benchmarking Machine-Generated Text Detection.” arXiv. https://doi.org/10.48550/ARXIV.2303.14822.
He, Xuehai, Diji Yang, Weixi Feng, Tsu-Jui Fu, Arjun Akula, Varun Jampani, Pradyumna Narayana, Sugato Basu, William Yang Wang, and Xin Eric Wang. 2022. “CPL: Counterfactual Prompt Learning for Vision and Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2210.10362.
He, Yanzhang, Tara N. Sainath, Rohit Prabhavalkar, Ian McGraw, Raziel Alvarez, Ding Zhao, David Rybach, et al. 2018. “Streaming End-to-End Speech Recognition for Mobile Devices.” arXiv. https://doi.org/10.48550/ARXIV.1811.06621.
He, Yihui, Xiangyu Zhang, and Jian Sun. 2017. “Channel Pruning for Accelerating Very Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1707.06168.
He, Zhenliang, Meina Kan, and Shiguang Shan. 2021. “EigenGAN: Layer-Wise Eigen-Learning for GANs.” arXiv. https://doi.org/10.48550/ARXIV.2104.12476.
He, Zhezhi, Boqing Gong, and Deliang Fan. 2018. “Optimize Deep Convolutional Neural Network with Ternarized Weights and High Accuracy.” arXiv. https://doi.org/10.48550/ARXIV.1807.07948.
Heaton, J. B., N. G. Polson, and J. H. Witte. 2016. “Deep Learning in Finance.” arXiv. https://doi.org/10.48550/ARXIV.1602.06561.
Heaton, Jeff. 2020. “Applications of Deep Neural Networks with Keras.” arXiv. https://doi.org/10.48550/ARXIV.2009.05673.
Heckerman, David. 2020. “A Tutorial on Learning with Bayesian Networks.” arXiv. https://doi.org/10.48550/ARXIV.2002.00269.
Hegre, Håvard, Nils W Metternich, Håvard Mokleiv Nygård, and Julian Wucherpfennig. 2017. “Introduction.” Journal of Peace Research 54 (February). https://doi.org/10.1177/0022343317691330.
Heinrich, Johannes, and David Silver. 2016. “Deep Reinforcement Learning from Self-Play in Imperfect-Information Games.” arXiv. https://doi.org/10.48550/ARXIV.1603.01121.
Heinson, A. P. 2008. “Single Top Quarks at the Tevatron.” arXiv. https://doi.org/10.48550/ARXIV.0809.0960.
Heist, Nicolas, and Heiko Paulheim. 2019. “Uncovering the Semantics of Wikipedia Categories.” arXiv. https://doi.org/10.48550/ARXIV.1906.12089.
Hellendoorn, Vincent J., and Premkumar Devanbu. 2017. “Are Deep Neural Networks the Best Choice for Modeling Source Code?” Proceedings of the 2017 11th Joint Meeting on Foundations of Software Engineering, August. https://doi.org/10.1145/3106237.3106290.
Helwegen, Koen, James Widdicombe, Lukas Geiger, Zechun Liu, Kwang-Ting Cheng, and Roeland Nusselder. 2019. “Latent Weights Do Not Exist: Rethinking Binarized Neural Network Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1906.02107.
Henaff, Mikael, Joan Bruna, and Yann LeCun. 2015. “Deep Convolutional Networks on Graph-Structured Data.” arXiv. https://doi.org/10.48550/ARXIV.1506.05163.
Henderson, John. 2002. “Knowing Someone Through Their Books: Pliny on Uncle Pliny ("Epistles" 3.5).” Classical Philology 97 (July). https://doi.org/10.1086/449587.
Hendricks, Lisa Anne, John Mellor, Rosalia Schneider, Jean-Baptiste Alayrac, and Aida Nematzadeh. 2021. “Decoupling the Role of Data, Attention, and Losses in Multimodal Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2102.00529.
Hendrycks, Dan, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, Eric Tang, Dawn Song, and Jacob Steinhardt. 2021. “Measuring Mathematical Problem Solving with the MATH Dataset.” arXiv. https://doi.org/10.48550/ARXIV.2103.03874.
Hendrycks, Dan, and Thomas Dietterich. 2019. “Benchmarking Neural Network Robustness to Common Corruptions and Perturbations.” arXiv. https://doi.org/10.48550/ARXIV.1903.12261.
Hendrycks, Dan, Kevin Zhao, Steven Basart, Jacob Steinhardt, and Dawn Song. 2019. “Natural Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1907.07174.
Hendy, Amr, Mohamed Abdelrehim, Amr Sharaf, Vikas Raunak, Mohamed Gabr, Hitokazu Matsushita, Young Jin Kim, Mohamed Afify, and Hany Hassan Awadalla. 2023. “How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation.” arXiv. https://doi.org/10.48550/ARXIV.2302.09210.
Hensman, James, Nicolo Fusi, and Neil D. Lawrence. 2013. “Gaussian Processes for Big Data.” arXiv. https://doi.org/10.48550/ARXIV.1309.6835.
Hensman, James, and Neil D. Lawrence. 2014. “Nested Variational Compression in Deep Gaussian Processes.” arXiv. https://doi.org/10.48550/ARXIV.1412.1370.
Henzinger, Thomas A., Mathias Lechner, and Đorđe Žikelić. 2020. “Scalable Verification of Quantized Neural Networks (Technical Report).” arXiv. https://doi.org/10.48550/ARXIV.2012.08185.
Herath, Sachini, Saghar Irandoust, Bowen Chen, Yiming Qian, Pyojin Kim, and Yasutaka Furukawa. 2021. “Fusion-DHL: WiFi, IMU, and Floorplan Fusion for Dense History of Locations in Indoor Environments.” arXiv. https://doi.org/10.48550/ARXIV.2105.08837.
Herce-Zelaya, Julio, Carlos Porcel, Álvaro Tejeda-Lorente, Juan Bernabé-Moreno, and Enrique Herrera-Viedma. 2022. “Introducing CSP Dataset: A Dataset Optimized for the Study of the Cold Start Problem in Recommender Systems.” Information 14 (December). https://doi.org/10.3390/info14010019.
Hermann, Karl Moritz, Felix Hill, Simon Green, Fumin Wang, Ryan Faulkner, Hubert Soyer, David Szepesvari, et al. 2017. “Grounded Language Learning in a Simulated 3D World.” arXiv. https://doi.org/10.48550/ARXIV.1706.06551.
Herzig, Jonathan, Pawel Krzysztof Nowak, Thomas Müller, Francesco Piccinno, and Julian Eisenschlos. 2020. “TaPas: Weakly Supervised Table Parsing via Pre-Training.” Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/2020.acl-main.398.
Hester, Todd, Matej Vecerik, Olivier Pietquin, Marc Lanctot, Tom Schaul, Bilal Piot, Dan Horgan, et al. 2017. “Deep q-Learning from Demonstrations.” arXiv. https://doi.org/10.48550/ARXIV.1704.03732.
Hewlett, Daniel, Alexandre Lacoste, Llion Jones, Illia Polosukhin, Andrew Fandrianto, Jay Han, Matthew Kelcey, and David Berthelot. 2016. “WikiReading: A Novel Large-Scale Language Understanding Task over Wikipedia.” arXiv. https://doi.org/10.48550/ARXIV.1608.03542.
Hidasi, Balázs, Alexandros Karatzoglou, Linas Baltrunas, and Domonkos Tikk. 2015. “Session-Based Recommendations with Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.06939.
Hildebrandt, Marcel, Jorge Andres Quintero Serna, Yunpu Ma, Martin Ringsquandl, Mitchell Joblin, and Volker Tresp. 2020. “Debate Dynamics for Human-Comprehensible Fact-Checking on Knowledge Graphs.” arXiv. https://doi.org/10.48550/ARXIV.2001.03436.
Hill, Felix, Antoine Bordes, Sumit Chopra, and Jason Weston. 2015. “The Goldilocks Principle: Reading Children’s Books with Explicit Memory Representations.” arXiv. https://doi.org/10.48550/ARXIV.1511.02301.
Hill, Felix, Kyunghyun Cho, Sébastien Jean, and Yoshua Bengio. 2017. “The Representational Geometry of Word Meanings Acquired by Neural Machine Translation Models.” Machine Translation 31 (April). https://doi.org/10.1007/s10590-017-9194-2.
Himmelstein, Daniel S., Antoine Lizee, Christine Hessler, Leo Brueggeman, Sabrina L. Chen, Dexter Hadley, Ari Green, Pouya Khankhanian, and Sergio E. Baranzini. 2016. “Systematic Integration of Biomedical Knowledge Prioritizes Drugs for Repurposing,” November. https://doi.org/10.1101/087619.
Hinton, G. E., and R. R. Salakhutdinov. 2006. “Reducing the Dimensionality of Data with Neural Networks.” Science 313 (July). https://doi.org/10.1126/science.1127647.
Hinton, Geoffrey E., Nitish Srivastava, Alex Krizhevsky, Ilya Sutskever, and Ruslan R. Salakhutdinov. 2012. “Improving Neural Networks by Preventing Co-Adaptation of Feature Detectors.” arXiv. https://doi.org/10.48550/ARXIV.1207.0580.
Hinton, Geoffrey, Li Deng, Dong Yu, George Dahl, Abdel-rahman Mohamed, Navdeep Jaitly, Andrew Senior, et al. 2012. “Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups.” IEEE Signal Processing Magazine 29 (November). https://doi.org/10.1109/msp.2012.2205597.
Hinton, Geoffrey, Oriol Vinyals, and Jeff Dean. 2015. “Distilling the Knowledge in a Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.1503.02531.
Hirn, Matthew, Nicolas Poilvert, and Stéphane Mallat. 2015. “Quantum Energy Regression Using Scattering Transforms.” arXiv. https://doi.org/10.48550/ARXIV.1502.02077.
Hirosawa, Takanobu, Yukinori Harada, Masashi Yokose, Tetsu Sakamoto, Ren Kawamura, and Taro Shimizu. 2023. “Diagnostic Accuracy of Differential-Diagnosis Lists Generated by Generative Pretrained Transformer 3 Chatbot for Clinical Vignettes with Common Chief Complaints: A Pilot Study.” International Journal of Environmental Research and Public Health 20 (February). https://doi.org/10.3390/ijerph20043378.
Hisadome, Yoichiro, and Yusuke Matsui. 2021. “Cascading Feature Extraction for Fast Point Cloud Registration.” arXiv. https://doi.org/10.48550/ARXIV.2110.12204.
Hitaj, Briland, Paolo Gasti, Giuseppe Ateniese, and Fernando Perez-Cruz. 2017. “PassGAN: A Deep Learning Approach for Password Guessing.” arXiv. https://doi.org/10.48550/ARXIV.1709.00440.
Ho, Jonathan, and Stefano Ermon. 2016. “Generative Adversarial Imitation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.03476.
Hobert, James P. 2011. “The Data Augmentation Algorithm: Theory and Methodology.” Handbook of Markov Chain Monte Carlo, May. https://doi.org/10.1201/b10905-11.
Hochreiter, Sepp. 1998. “The Vanishing Gradient Problem During Learning Recurrent Neural Nets and Problem Solutions.” International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 06 (April). https://doi.org/10.1142/s0218488598000094.
Hochreiter, Sepp, and Jürgen Schmidhuber. 1997. “Long Short-Term Memory.” Neural Computation 9 (November). https://doi.org/10.1162/neco.1997.9.8.1735.
Hoecker, A., P. Speckmayer, J. Stelzer, J. Therhaag, E. von Toerne, H. Voss, M. Backes, et al. 2007. “TMVA - Toolkit for Multivariate Data Analysis.” arXiv. https://doi.org/10.48550/ARXIV.PHYSICS/0703039.
Hoffer, Elad, Itay Hubara, and Daniel Soudry. 2017. “Train Longer, Generalize Better: Closing the Generalization Gap in Large Batch Training of Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1705.08741.
Hoffman, Judy, Eric Tzeng, Taesung Park, Jun-Yan Zhu, Phillip Isola, Kate Saenko, Alexei A. Efros, and Trevor Darrell. 2017. “CyCADA: Cycle-Consistent Adversarial Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1711.03213.
Hoffman, Matt, David M. Blei, Chong Wang, and John Paisley. 2012. “Stochastic Variational Inference.” arXiv. https://doi.org/10.48550/ARXIV.1206.7051.
Hoffman, Matthew D., and Andrew Gelman. 2011. “The No-u-Turn Sampler: Adaptively Setting Path Lengths in Hamiltonian Monte Carlo.” arXiv. https://doi.org/10.48550/ARXIV.1111.4246.
Hofherr, Florian, Lukas Koestler, Florian Bernard, and Daniel Cremers. 2022. “Neural Implicit Representations for Physical Parameter Inference from a Single Video.” arXiv. https://doi.org/10.48550/ARXIV.2204.14030.
Hofstätter, Sebastian, Sheng-Chieh Lin, Jheng-Hong Yang, Jimmy Lin, and Allan Hanbury. 2021. “Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware Sampling.” Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3404835.3462891.
Hojjati, Hadi, Thi Kieu Khanh Ho, and Narges Armanfard. 2024. “Self-Supervised Anomaly Detection in Computer Vision and Beyond: A Survey and Outlook.” Neural Networks 172 (April). https://doi.org/10.1016/j.neunet.2024.106106.
Holden, Daniel, Taku Komura, and Jun Saito. 2017. “Phase-Functioned Neural Networks for Character Control.” ACM Transactions on Graphics 36 (July). https://doi.org/10.1145/3072959.3073663.
Hollmann, Noah, Samuel Müller, and Frank Hutter. 2023. “Large Language Models for Automated Data Science: Introducing CAAFE for Context-Aware Automated Feature Engineering.” arXiv. https://doi.org/10.48550/ARXIV.2305.03403.
Holmes, Jason, Zhengliang Liu, Lian Zhang, Yuzhen Ding, Terence T. Sio, Lisa A. McGee, Jonathan B. Ashman, et al. 2023. “Evaluating Large Language Models on a Highly-Specialized Topic, Radiation Oncology Physics.” Frontiers in Oncology 13 (July). https://doi.org/10.3389/fonc.2023.1219326.
Holroyd, Alexander E. 2003. “Sharp Metastability Threshold for Two-Dimensional Bootstrap Percolation.” Probability Theory and Related Fields 125 (February). https://doi.org/10.1007/s00440-002-0239-x.
Holzinger, Andreas, Chris Biemann, Constantinos S. Pattichis, and Douglas B. Kell. 2017. “What Do We Need to Build Explainable AI Systems for the Medical Domain?” arXiv. https://doi.org/10.48550/ARXIV.1712.09923.
Homayounpour, M. Mehdi, and Mohammad Moattar. 2009. “A Simple but Efficient Real-Time Voice Activity Detection Algorithm.” Zenodo, August. https://doi.org/10.5281/ZENODO.41799.
Hong, Yining, Haoyu Zhen, Peihao Chen, Shuhong Zheng, Yilun Du, Zhenfang Chen, and Chuang Gan. 2023. “3D-LLM: Injecting the 3D World into Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2307.12981.
Hooi, Bryan, Hyun Ah Song, Alex Beutel, Neil Shah, Kijung Shin, and Christos Faloutsos. 2016. “FRAUDAR.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939747.
Hori, Takaaki, Shinji Watanabe, Yu Zhang, and William Chan. 2017. “Advances in Joint CTC-Attention Based End-to-End Speech Recognition with a Deep CNN Encoder and RNN-LM.” arXiv. https://doi.org/10.48550/ARXIV.1706.02737.
Horowitz, Mark. 2014. “1.1 Computing’s Energy Problem (and What We Can Do about It).” 2014 IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC), February. https://doi.org/10.1109/isscc.2014.6757323.
Horvath, Samuel, Chen-Yu Ho, Ludovit Horvath, Atal Narayan Sahu, Marco Canini, and Peter Richtarik. 2019. “Natural Compression for Distributed Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1905.10988.
Horvitz, Eric, and Deirdre Mulligan. 2015. “Data, Privacy, and the Greater Good.” Science 349 (July). https://doi.org/10.1126/science.aac4520.
Hoshen, Yedid. 2017. “VAIN: Attentional Multi-Agent Predictive Modeling.” arXiv. https://doi.org/10.48550/ARXIV.1706.06122.
Hosoya, Yuzo. 1991. “The Decomposition and Measurement of the Interdependency Between Second-Order Stationary Processes.” Probability Theory and Related Fields 88 (December). https://doi.org/10.1007/bf01192551.
Hospedales, Timothy, Antreas Antoniou, Paul Micaelli, and Amos Storkey. 2020. “Meta-Learning in Neural Networks: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2004.05439.
Hosseini-Asl, Ehsan, Georgy Gimel’farb, and Ayman El-Baz. 2016. “Alzheimer’s Disease Diagnostics by a Deeply Supervised Adaptable 3D Convolutional Network.” arXiv. https://doi.org/10.48550/ARXIV.1607.00556.
Hosten, Benoît, Chadi Abbara, Benoît Petit, Angélique Dauvin, Fanchon Bourasset, Robert Farinotti, Patrick Gonin, and Laurence Bonhomme-Faivre. 2008. “Effect of Interleukin-2 Pretreatment on Paclitaxel Absorption and Tissue Disposition After Oral and Intravenous Administration in Mice.” Drug Metabolism and Disposition 36 (May). https://doi.org/10.1124/dmd.107.019091.
Hou, Haodi, Jing Huo, and Yang Gao. 2018. “Cross-Domain Adversarial Auto-Encoder.” arXiv. https://doi.org/10.48550/ARXIV.1804.06078.
Hou, Lu, Zhiqi Huang, Lifeng Shang, Xin Jiang, Xiao Chen, and Qun Liu. 2020. “DynaBERT: Dynamic BERT with Adaptive Width and Depth.” arXiv. https://doi.org/10.48550/ARXIV.2004.04037.
Hou, Lu, and James T. Kwok. 2018. “Loss-Aware Weight Quantization of Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1802.08635.
Hou, Lu, Quanming Yao, and James T. Kwok. 2016. “Loss-Aware Binarization of Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.01600.
Hou, Min, Chang Xu, Zhi Li, Yang Liu, Weiqing Liu, Enhong Chen, and Jiang Bian. 2022. “Multi-Granularity Residual Learning with Confidence Estimation for Time Series Prediction.” Proceedings of the ACM Web Conference 2022, April. https://doi.org/10.1145/3485447.3512056.
Hou, Yupeng, Zhankui He, Julian McAuley, and Wayne Xin Zhao. 2022. “Learning Vector-Quantized Item Representation for Transferable Sequential Recommenders.” arXiv. https://doi.org/10.48550/ARXIV.2210.12316.
Houthooft, Rein, Xi Chen, Yan Duan, John Schulman, Filip De Turck, and Pieter Abbeel. 2016. “VIME: Variational Information Maximizing Exploration.” arXiv. https://doi.org/10.48550/ARXIV.1605.09674.
Howard, Andrew G., Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.” arXiv. https://doi.org/10.48550/ARXIV.1704.04861.
Hrinchuk, Oleksii, Mariya Popova, and Boris Ginsburg. 2019. “Correction of Automatic Speech Recognition with Transformer Sequence-to-Sequence Model.” arXiv. https://doi.org/10.48550/ARXIV.1910.10697.
Hsieh, Cheng-Yu, Chun-Liang Li, Chih-Kuan Yeh, Hootan Nakhost, Yasuhisa Fujii, Alexander Ratner, Ranjay Krishna, Chen-Yu Lee, and Tomas Pfister. 2023. “Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes.” arXiv. https://doi.org/10.48550/ARXIV.2305.02301.
Hsieh, Tsung-Yu, Yiwei Sun, Xianfeng Tang, Suhang Wang, and Vasant G. Honavar. 2021. “SrVARM: State Regularized Vector Autoregressive Model for Joint Learning of Hidden State Transitions and State-Dependent Inter-Variable Dependencies from Multi-Variate Time Series.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3450116.
Hsu, Daniel, and Sham M. Kakade. 2012. “Learning Mixtures of Spherical Gaussians: Moment Methods and Spectral Decompositions.” arXiv. https://doi.org/10.48550/ARXIV.1206.5766.
Hsu, Kyle, Sergey Levine, and Chelsea Finn. 2018. “Unsupervised Learning via Meta-Learning.” arXiv. https://doi.org/10.48550/ARXIV.1810.02334.
Hu, Binbin, Zhiqiang Zhang, Chuan Shi, Jun Zhou, Xiaolong Li, and Yuan Qi. 2019. “Cash-Out User Detection Based on Attributed Heterogeneous Information Network with a Hierarchical Attention Mechanism.” Proceedings of the AAAI Conference on Artificial Intelligence 33 (July). https://doi.org/10.1609/aaai.v33i01.3301946.
Hu, Chuanbo, Minglei Yin, Bin Liu, Xin Li, and Yanfang Ye. 2021. “Detection of Illicit Drug Trafficking Events on Instagram.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3481908.
Hu, Edward J., Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2021. “LoRA: Low-Rank Adaptation of Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2106.09685.
Hu, Fenyu, Yanqiao Zhu, Shu Wu, Weiran Huang, Liang Wang, and Tieniu Tan. 2021. “GraphAIR: Graph Representation Learning with Neighborhood Aggregation and Interaction.” Pattern Recognition 112 (April). https://doi.org/10.1016/j.patcog.2020.107745.
Hu, Guanyu, Junxian Geng, Yishu Xue, and Huiyan Sang. 2023. “Bayesian Spatial Homogeneity Pursuit of Functional Data: An Application to the u.s. Income Distribution.” Bayesian Analysis 18 (June). https://doi.org/10.1214/22-ba1320.
Hu, Hanzhang, Debadeepta Dey, Martial Hebert, and J. Andrew Bagnell. 2017. “Learning Anytime Predictions in Neural Networks via Adaptive Loss Balancing.” arXiv. https://doi.org/10.48550/ARXIV.1708.06832.
Hu, Hongyu, Tiancheng Lin, Jie Wang, Zhenbang Sun, and Yi Xu. 2023. “Context-Aware Prompt Tuning for Vision-Language Model with Dual-Alignment.” arXiv. https://doi.org/10.48550/ARXIV.2309.04158.
Hu, Jie, Li Shen, Samuel Albanie, Gang Sun, and Enhua Wu. 2017. “Squeeze-and-Excitation Networks.” arXiv. https://doi.org/10.48550/ARXIV.1709.01507.
Hu, Jinlong, Junjie Liang, and Shoubin Dong. 2017. “iBGP: A Bipartite Graph Propagation Approach for Mobile Advertising Fraud Detection.” Mobile Information Systems 2017. https://doi.org/10.1155/2017/6412521.
Hu, Mingzhe, Shaoyan Pan, Yuheng Li, and Xiaofeng Yang. 2023. “Advancing Medical Imaging with Language Models: A Journey from n-Grams to ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2304.04920.
Hu, Shengchao, Li Shen, Ya Zhang, and Dacheng Tao. 2023. “Prompt-Tuning Decision Transformer with Preference Ranking.” arXiv. https://doi.org/10.48550/ARXIV.2305.09648.
Hu, Sihao, Zhen Zhang, Bingqiao Luo, Shengliang Lu, Bingsheng He, and Ling Liu. 2023. “BERT4ETH: A Pre-Trained Transformer for Ethereum Fraud Detection.” Proceedings of the ACM Web Conference 2023, April. https://doi.org/10.1145/3543507.3583345.
Hu, Ting, Christoph Meinel, and Haojin Yang. 2023. “Scaled Prompt-Tuning for Few-Shot Natural Language Generation.” arXiv. https://doi.org/10.48550/ARXIV.2309.06759.
Hu, Xinxin, Haotian Chen, Junjie Zhang, Hongchang Chen, Shuxin Liu, Xing Li, Yahui Wang, and Xiangyang Xue. 2023. “GAT-COBO: Cost-Sensitive Graph Neural Network for Telecom Fraud Detection.” arXiv. https://doi.org/10.48550/ARXIV.2303.17334.
Hu, Yan, Qingyu Chen, Jingcheng Du, Xueqing Peng, Vipina Kuttichi Keloth, Xu Zuo, Yujia Zhou, et al. 2023. “Improving Large Language Models for Clinical Named Entity Recognition via Prompt Engineering.” arXiv. https://doi.org/10.48550/ARXIV.2303.16416.
Hu, Yiming, Siyang Sun, Jianquan Li, Jiagang Zhu, Xingang Wang, and Qingyi Gu. 2019. “Multi-Loss-Aware Channel Pruning of Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1902.10364.
Hu, Y., and S. Peng. 1995. “Solution of Forward-Backward Stochastic Differential Equations.” Probability Theory and Related Fields 103 (June). https://doi.org/10.1007/bf01204218.
Hu, Yujing, Qing Da, Anxiang Zeng, Yang Yu, and Yinghui Xu. 2018. “Reinforcement Learning to Rank in e-Commerce Search Engine.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3219846.
Hu, Zhiyuan, Yunsheng Li, Jiancheng Lyu, Dashan Gao, and Nuno Vasconcelos. 2023. “Dense Network Expansion for Class Incremental Learning.” arXiv. https://doi.org/10.48550/ARXIV.2303.12696.
Hu, Ziniu, Ahmet Iscen, Chen Sun, Zirui Wang, Kai-Wei Chang, Yizhou Sun, Cordelia Schmid, David A. Ross, and Alireza Fathi. 2022. “REVEAL: Retrieval-Augmented Visual-Language Pre-Training with Multi-Source Multimodal Knowledge Memory.” arXiv. https://doi.org/10.48550/ARXIV.2212.05221.
Hua, William, Hongyuan Mei, Sarah Zohar, Magali Giral, and Yanxun Xu. 2022. “Personalized Dynamic Treatment Regimes in Continuous Time: A Bayesian Approach for Optimizing Clinical Decisions with Timing.” Bayesian Analysis 17 (September). https://doi.org/10.1214/21-ba1276.
Huang, Allen H., Hui Wang, and Yi Yang. 2023. “<Scp>FinBERT</Scp>: A Large Language Model for Extracting Information from Financial Text*.” Contemporary Accounting Research 40 (January). https://doi.org/10.1111/1911-3846.12832.
Huang, Binxuan, and Kathleen M. Carley. 2019. “Parameterized Convolutional Neural Networks for Aspect Level Sentiment Classification.” arXiv. https://doi.org/10.48550/ARXIV.1909.06276.
Huang, Calista, Alyssa Ma, Suchir Vyasamudri, Eugenie Puype, Sayem Kamal, Juan Belza Garcia, Salar Cheema, and Michael Lutz. 2024. “ACCESS: Prompt Engineering for Automated Web Accessibility Violation Corrections.” arXiv. https://doi.org/10.48550/ARXIV.2401.16450.
Huang, Chenghao, Yanbo Cao, Yinlong Wen, Tao Zhou, and Yanru Zhang. 2024. “PokerGPT: An End-to-End Lightweight Solver for Multi-Player Texas Hold’em via Large Language Model.” arXiv. https://doi.org/10.48550/ARXIV.2401.06781.
Huang, Di, Jacob Bartel, and John Palowitch. 2021. “Recurrent Graph Neural Networks for Rumor Detection in Online Forums.” arXiv. https://doi.org/10.48550/ARXIV.2108.03548.
Huang, Gao, Shichen Liu, Laurens van der Maaten, and Kilian Q. Weinberger. 2017. “CondenseNet: An Efficient DenseNet Using Learned Group Convolutions.” arXiv. https://doi.org/10.48550/ARXIV.1711.09224.
Huang, Gao, Zhuang Liu, Laurens van der Maaten, and Kilian Q. Weinberger. 2016. “Densely Connected Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1608.06993.
Huang, Hsin-Yuan, Chenguang Zhu, Yelong Shen, and Weizhu Chen. 2017. “FusionNet: Fusing via Fully-Aware Attention with Application to Machine Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.1711.07341.
Huang, Jialu, Jing Liao, and Sam Kwong. 2020. “Unsupervised Image-to-Image Translation via Pre-Trained StyleGAN2 Network.” arXiv. https://doi.org/10.48550/ARXIV.2010.05713.
Huang, Jiaxing, Jingyi Zhang, Han Qiu, Sheng Jin, and Shijian Lu. 2023. “Prompt Ensemble Self-Training for Open-Vocabulary Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.2306.16658.
Huang, Jonathan, Vivek Rathod, Chen Sun, Menglong Zhu, Anoop Korattikara, Alireza Fathi, Ian Fischer, et al. 2016. “Speed/Accuracy Trade-Offs for Modern Convolutional Object Detectors.” arXiv. https://doi.org/10.48550/ARXIV.1611.10012.
Huang, Jui-Ting, Ashish Sharma, Shuying Sun, Li Xia, David Zhang, Philip Pronin, Janani Padmanabhan, Giuseppe Ottaviano, and Linjun Yang. 2020. “Embedding-Based Retrieval in Facebook Search.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403305.
Huang, Lianzhe, Dehong Ma, Sujian Li, Xiaodong Zhang, and Houfeng WANG. 2019. “Text Level Graph Neural Network for Text Classification.” arXiv. https://doi.org/10.48550/ARXIV.1910.02356.
Huang, Lifu, Heng Ji, Kyunghyun Cho, and Clare R. Voss. 2017. “Zero-Shot Transfer Learning for Event Extraction,” July. http://arxiv.org/abs/1707.01066v1.
Huang, Li, and Dianbo Liu. 2019. “Patient Clustering Improves Efficiency of Federated Machine Learning to Predict Mortality and Hospital Stay Time Using Distributed Electronic Medical Records.” arXiv. https://doi.org/10.48550/ARXIV.1903.09296.
Huang, Mengda, Yang Liu, Xiang Ao, Kuan Li, Jianfeng Chi, Jinghua Feng, Hao Yang, and Qing He. 2022. “AUC-Oriented Graph Neural Network for Fraud Detection.” Proceedings of the ACM Web Conference 2022, April. https://doi.org/10.1145/3485447.3512178.
Huang, Mingkun, Meng Cai, Jun Zhang, Yang Zhang, Yongbin You, Yi He, and Zejun Ma. 2020. “Dynamic Latency Speech Recognition with Asynchronous Revision.” arXiv. https://doi.org/10.48550/ARXIV.2011.01570.
Huang, Qiang, Makoto Yamada, Yuan Tian, Dinesh Singh, Dawei Yin, and Yi Chang. 2020. “GraphLIME: Local Interpretable Model Explanations for Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2001.06216.
Huang, Qingqing, Daniel S. Park, Tao Wang, Timo I. Denk, Andy Ly, Nanxin Chen, Zhengdong Zhang, et al. 2023. “Noise2Music: Text-Conditioned Music Generation with Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.03917.
Huang, Qing, Yanbang Sun, Zhenchang Xing, Min Yu, Xiwei Xu, and Qinghua Lu. 2023. “API Entity and Relation Joint Extraction from Text via Dynamic Prompt-Tuned Language Model.” arXiv. https://doi.org/10.48550/ARXIV.2301.03987.
Huang, Qing, Zhiqiang Yuan, Zhenchang Xing, Xiwei Xu, Liming Zhu, and Qinghua Lu. 2022. “Prompt-Tuned Code Language Model as a Neural Knowledge Base for Type Inference in Statically-Typed Partial Code.” arXiv. https://doi.org/10.48550/ARXIV.2208.05361.
Huang, Qiuhua, Renke Huang, Weituo Hao, Jie Tan, Rui Fan, and Zhenyu Huang. 2019. “Adaptive Power System Emergency Control Using Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1903.03712.
Huang, Rongjie, Chenye Cui, Feiyang Chen, Yi Ren, Jinglin Liu, Zhou Zhao, Baoxing Huai, and Zhefeng Wang. 2021. “SingGAN: Generative Adversarial Network for High-Fidelity Singing Voice Generation.” arXiv. https://doi.org/10.48550/ARXIV.2110.07468.
Huang, Rongjie, Jiawei Huang, Dongchao Yang, Yi Ren, Luping Liu, Mingze Li, Zhenhui Ye, Jinglin Liu, Xiang Yin, and Zhou Zhao. 2023. “Make-an-Audio: Text-to-Audio Generation with Prompt-Enhanced Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.12661.
Huang, Rongjie, Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu, Yi Ren, and Zhou Zhao. 2022. “FastDiff: A Fast Conditional Diffusion Model for High-Quality Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.2204.09934.
Huang, Shaohan, Li Dong, Wenhui Wang, Yaru Hao, Saksham Singhal, Shuming Ma, Tengchao Lv, et al. 2023. “Language Is Not All You Need: Aligning Perception with Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.14045.
Huang, Sicong, Asli Celikyilmaz, and Haoran Li. 2022. “ED-FAITH: Evaluating Dialogue Summarization on Faithfulness.” arXiv. https://doi.org/10.48550/ARXIV.2211.08464.
Huang, Wenke, Mang Ye, Zekun Shi, He Li, and Bo Du. 2023. “Rethinking Federated Learning with Domain Shift: A Prototype View.” 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52729.2023.01565.
Huang, Wenlong, Pieter Abbeel, Deepak Pathak, and Igor Mordatch. 2022. “Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents.” arXiv. https://doi.org/10.48550/ARXIV.2201.07207.
Huang, Wenlong, Fei Xia, Dhruv Shah, Danny Driess, Andy Zeng, Yao Lu, Pete Florence, et al. 2023. “Grounded Decoding: Guiding Text Generation with Grounded Models for Embodied Agents.” arXiv. https://doi.org/10.48550/ARXIV.2303.00855.
Huang, Xuanwen, Yang Yang, Yang Wang, Chunping Wang, Zhisheng Zhang, Jiarong Xu, Lei Chen, and Michalis Vazirgiannis. 2022. “DGraph: A Large-Scale Financial Dataset for Graph Anomaly Detection.” arXiv. https://doi.org/10.48550/ARXIV.2207.03579.
Huang, Yaomin, Ning Liu, Zhengping Che, Zhiyuan Xu, Chaomin Shen, Yaxin Peng, Guixu Zhang, Xinmei Liu, Feifei Feng, and Jian Tang. 2023. “CP\(^3\): Channel Pruning Plug-in for Point-Based Networks.” arXiv. https://doi.org/10.48550/ARXIV.2303.13097.
Huang, Yihong, Liping Wang, Fan Zhang, and Xuemin Lin. 2022. “Unsupervised Graph Outlier Detection: Problem Revisit, New Insight, and Superior Method.” arXiv. https://doi.org/10.48550/ARXIV.2210.12941.
Huang, Yimin, Yujun Li, Hanrong Ye, Zhenguo Li, and Zhihua Zhang. 2020. “An Asymptotically Optimal Multi-Armed Bandit Algorithm and Hyperparameter Optimization.” arXiv. https://doi.org/10.48550/ARXIV.2007.05670.
Huang, Yipo, Quan Yuan, Xiangfei Sheng, Zhichao Yang, Haoning Wu, Pengfei Chen, Yuzhe Yang, Leida Li, and Weisi Lin. 2024. “AesBench: An Expert Benchmark for Multimodal Large Language Models on Image Aesthetics Perception.” arXiv. https://doi.org/10.48550/ARXIV.2401.08276.
Huang, Zehao, and Naiyan Wang. 2017a. “Data-Driven Sparse Structure Selection for Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1707.01213.
———. 2017b. “Like What You Like: Knowledge Distill via Neuron Selectivity Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1707.01219.
Huang, Zeyi, Haohan Wang, Eric P. Xing, and Dong Huang. 2020. “Self-Challenging Improves Cross-Domain Generalization.” arXiv. https://doi.org/10.48550/ARXIV.2007.02454.
Huang, Zhengjie, Yunyang Huang, Peng Qian, Jianhai Chen, and Qinming He. 2022. “Demystifying Bitcoin Address Behavior via Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2211.14582.
Huang, Zhening, Xiaoyang Wu, Hengshuang Zhao, Lei Zhu, Shujun Wang, Georgios Hadjidemetriou, and Ioannis Brilakis. 2023. “GeoSpark: Sparking up Point Cloud Segmentation with Geometry Clue.” arXiv. https://doi.org/10.48550/ARXIV.2303.08274.
Huang, Zhiheng, Wei Xu, and Kai Yu. 2015. “Bidirectional LSTM-CRF Models for Sequence Tagging.” arXiv. https://doi.org/10.48550/ARXIV.1508.01991.
Huang, Zhipeng, and Nikos Mamoulis. 2017. “Heterogeneous Information Network Embedding for Meta Path Based Proximity.” arXiv. https://doi.org/10.48550/ARXIV.1701.05291.
Hubara, Itay, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. 2016. “Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations.” arXiv. https://doi.org/10.48550/ARXIV.1609.07061.
Huber, Peter J. 1964. “Robust Estimation of a Location Parameter.” The Annals of Mathematical Statistics 35 (March). https://doi.org/10.1214/aoms/1177703732.
Hubert, Nicolas, Heiko Paulheim, Pierre Monnin, Armelle Brun, and Davy Monticolo. 2023. “Schema First! Learn Versatile Knowledge Graph Embeddings by Capturing Semantics with MASCHInE.” Proceedings of the 12th Knowledge Capture Conference 2023, December. https://doi.org/10.1145/3587259.3627550.
Huelsenbeck, John P., and Jonathan P. Bollback. 2001. “Empirical and Hierarchical Bayesian Estimation of Ancestral States.” Systematic Biology 50 (May). https://doi.org/10.1080/106351501300317978.
Hui, Francis K. C. 2016. “<Scp>boral</Scp> – Bayesian Ordination and Regression Analysis of Multivariate Abundance Data in <Scp>r</Scp>.” Methods in Ecology and Evolution 7 (January). https://doi.org/10.1111/2041-210x.12514.
Huling, Jared D., and Peter Z. G. Qian. 2018. “Fast Penalized Regression and Cross Validation for Tall Data with the Oem Package.” arXiv. https://doi.org/10.48550/ARXIV.1801.09661.
Humbert, Pierre, Julien Audiffren, Laurent Oudre, and Nicolas Vayatis. 2019. “Multivariate Convolutional Sparse Coding with Low Rank Tensor.” arXiv. https://doi.org/10.48550/ARXIV.1908.03367.
Hundman, Kyle, Valentino Constantinou, Christopher Laporte, Ian Colwell, and Tom Soderstrom. 2018. “Detecting Spacecraft Anomalies Using LSTMs and Nonparametric Dynamic Thresholding.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3219845.
Hunt, Tyler, Congzheng Song, Reza Shokri, Vitaly Shmatikov, and Emmett Witchel. 2018. “Chiron: Privacy-Preserving Machine Learning as a Service.” arXiv. https://doi.org/10.48550/ARXIV.1803.05961.
Hunter, Anthony. 2013. “A Probabilistic Approach to Modelling Uncertain Logical Arguments.” International Journal of Approximate Reasoning 54 (January). https://doi.org/10.1016/j.ijar.2012.08.003.
Hürlimann, Werner. 2015. Advances in Pure Mathematics 05. https://doi.org/10.4236/apm.2015.57039.
Hustad, Katherine C., Tristan J. Mahr, Phoebe Natzke, and Paul J. Rathouz. 2021. “Speech Development Between 30 and 119 Months in Typical Children i: Intelligibility Growth Curves for Single-Word and Multiword Productions.” Journal of Speech, Language, and Hearing Research 64 (October). https://doi.org/10.1044/2021_jslhr-21-00142.
Hutter, F., H. H. Hoos, K. Leyton-Brown, and T. Stuetzle. 2009. “ParamILS: An Automatic Algorithm Configuration Framework.” Journal of Artificial Intelligence Research 36 (October). https://doi.org/10.1613/jair.2861.
Hwang, Kyuyeon, and Wonyong Sung. 2015. “Online Sequence Training of Recurrent Neural Networks with Connectionist Temporal Classification.” arXiv. https://doi.org/10.48550/ARXIV.1511.06841.
Hwang, Min-Jae, Eunwoo Song, Ryuichi Yamamoto, Frank Soong, and Hong-Goo Kang. 2020. “Improving LPCNet-Based Text-to-Speech with Linear Prediction-Structured Mixture Density Network.” arXiv. https://doi.org/10.48550/ARXIV.2001.11686.
Hwang, Yohan, Jang Ho Lee, and Dongkwang Shin. 2023. “What Is Prompt Literacy? An Exploratory Study of Language Learners’ Development of New Literacy Skill Using Generative AI.” arXiv. https://doi.org/10.48550/ARXIV.2311.05373.
Iandola, Forrest N., Song Han, Matthew W. Moskewicz, Khalid Ashraf, William J. Dally, and Kurt Keutzer. 2016. “SqueezeNet: AlexNet-Level Accuracy with 50x Fewer Parameters and &Lt;0.5MB Model Size.” arXiv. https://doi.org/10.48550/ARXIV.1602.07360.
“ICASSP 2016.” 2015. IEEE Signal Processing Magazine 32 (July). https://doi.org/10.1109/msp.2015.2411564.
Ida, Yasutoshi, and Yasuhiro Fujiwara. 2019. “Network Implosion: Effective Model Compression for ResNets via Static Layer Pruning and Retraining.” arXiv. https://doi.org/10.48550/ARXIV.1906.03826.
Ildar, Rakhmatulin. 2021. “Machine Vision for Low-Cost Remote Control of Mosquitoes by Power Laser.” Journal of Real-Time Image Processing 18 (February). https://doi.org/10.1007/s11554-021-01079-x.
Ilhan, Fatih, Gong Su, and Ling Liu. 2023. “ScaleFL: Resource-Adaptive Federated Learning with Heterogeneous Clients.” 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52729.2023.02350.
Imai, Kosuke, and David A. van Dyk. 2005. “A Bayesian Analysis of the Multinomial Probit Model Using Marginal Data Augmentation.” Journal of Econometrics 124 (February). https://doi.org/10.1016/j.jeconom.2004.02.002.
Imai, Kosuke, and Michael Lingzhi Li. 2021. “Experimental Evaluation of Individualized Treatment Rules.” Journal of the American Statistical Association 118 (June). https://doi.org/10.1080/01621459.2021.1923511.
———. 2022. “Statistical Inference for Heterogeneous Treatment Effects Discovered by Generic Machine Learning in Randomized Experiments.” arXiv. https://doi.org/10.48550/ARXIV.2203.14511.
Imai, Kosuke, and Marc Ratkovic. 2013. “Estimating Treatment Effect Heterogeneity in Randomized Program Evaluation.” The Annals of Applied Statistics 7 (March). https://doi.org/10.1214/12-aoas593.
“Inaugural Image and Vision Computing Outstanding Young Researcher Award Winner Announced.” 2012. Image and Vision Computing 30 (September). https://doi.org/10.1016/j.imavis.2012.07.008.
“Independent Component Analysis and Signal Separation.” 2009. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-642-00599-2.
Ingle, R. Reeve, Yasuhisa Fujii, Thomas Deselaers, Jonathan Baccash, and Ashok C. Popat. 2019. “A Scalable Handwritten Text Recognition System.” arXiv. https://doi.org/10.48550/ARXIV.1904.09150.
Inoue, Naoto, Kotaro Kikuchi, Edgar Simo-Serra, Mayu Otani, and Kota Yamaguchi. 2023. “LayoutDM: Discrete Diffusion Model for Controllable Layout Generation.” arXiv. https://doi.org/10.48550/ARXIV.2303.08137.
Ioannidis, Vassilis N., Dimitris Berberidis, and Georgios B. Giannakis. 2019. “GraphSAC: Detecting Anomalies in Large-Scale Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1910.09589.
Ioannou, Yani, Duncan Robertson, Roberto Cipolla, and Antonio Criminisi. 2017. “Deep Roots: Improving CNN Efficiency with Hierarchical Filter Groups.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July. https://doi.org/10.1109/cvpr.2017.633.
Ioffe, Sergey, and Christian Szegedy. 2015. “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift.” arXiv. https://doi.org/10.48550/ARXIV.1502.03167.
Iscen, Ahmet, Yannis Avrithis, Giorgos Tolias, Teddy Furon, and Ondrej Chum. 2017. “Fast Spectral Ranking for Similarity Search.” arXiv. https://doi.org/10.48550/ARXIV.1703.06935.
Iscoe, I. 1986. “A Weighted Occupation Time for a Class of Measured-Valued Branching Processes.” Probability Theory and Related Fields 71 (January). https://doi.org/10.1007/bf00366274.
Ishihara, Tatsuma, and Daisuke Saito. 2020. “Attention-Based Speaker Embeddings for One-Shot Voice Conversion.” Interspeech 2020, October. https://doi.org/10.21437/interspeech.2020-2512.
Ismail, Aya Abdelsalam, Mohamed Gunady, Héctor Corrada Bravo, and Soheil Feizi. 2020. “Benchmarking Deep Learning Interpretability in Time Series Predictions.” arXiv. https://doi.org/10.48550/ARXIV.2010.13924.
Isola, Phillip, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. 2016. “Image-to-Image Translation with Conditional Adversarial Networks,” November. http://arxiv.org/abs/1611.07004v3.
Iten, Raban, Tony Metger, Henrik Wilming, Lídia del Rio, and Renato Renner. 2020. “Discovering Physical Concepts with Neural Networks.” Physical Review Letters 124 (January). https://doi.org/10.1103/physrevlett.124.010508.
Ivashchenko, Tetiana, Andrii Ivashchenko, and Nelia Vasylets. 2023. “THE WAYS OF INTRODUCING AI/ML-BASED PREDICTION METHODS FOR THE IMPROVEMENT OF THE SYSTEM OF GOVERNMENT SOCIO-ECONOMIC ADMINISTRATION IN UKRAINE.” Business: Theory and Practice 24 (November). https://doi.org/10.3846/btp.2023.18733.
Iyer, Rishabh, Stefanie Jegelka, and Jeff Bilmes. 2013. “Fast Semidifferential-Based Submodular Function Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1308.1006.
Jacob, Benoit, Skirmantas Kligys, Bo Chen, Menglong Zhu, Matthew Tang, Andrew Howard, Hartwig Adam, and Dmitry Kalenichenko. 2017. “Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference.” arXiv. https://doi.org/10.48550/ARXIV.1712.05877.
Jaderberg, Max, Andrea Vedaldi, and Andrew Zisserman. 2014. “Speeding up Convolutional Neural Networks with Low Rank Expansions.” arXiv. https://doi.org/10.48550/ARXIV.1405.3866.
Jaeger, Byron C., Nicholas J. Tierney, and Noah R. Simon. 2020. “When to Impute? Imputation Before and During Cross-Validation.” arXiv. https://doi.org/10.48550/ARXIV.2010.00718.
Jaegle, Andrew, Felix Gimeno, Andrew Brock, Andrew Zisserman, Oriol Vinyals, and Joao Carreira. 2021. “Perceiver: General Perception with Iterative Attention.” arXiv. https://doi.org/10.48550/ARXIV.2103.03206.
Jain, Himanshu, Venkatesh Balasubramanian, Bhanu Chunduri, and Manik Varma. 2019. “Slice.” Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, January. https://doi.org/10.1145/3289600.3290979.
Jain, Prachi, Sushant Rathi, and Soumen Chakrabarti. 2020a. “Knowledge Base Completion: Baseline Strikes Back (Again).” arXiv. https://doi.org/10.48550/ARXIV.2005.00804.
———. 2020b. “Temporal Knowledge Base Completion: New Algorithms and Evaluation Protocols.” arXiv. https://doi.org/10.48550/ARXIV.2005.05035.
Jakovetic, Dusan, Joao Xavier, and Jose M. F. Moura. 2011. “Fast Distributed Gradient Methods.” arXiv. https://doi.org/10.48550/ARXIV.1112.2972.
Jang, Eric, Shixiang Gu, and Ben Poole. 2016. “Categorical Reparameterization with Gumbel-Softmax.” arXiv. https://doi.org/10.48550/ARXIV.1611.01144.
Jang, Jun-Woo, Sehwan Lee, Dongyoung Kim, Hyunsun Park, Ali Shafiee Ardestani, Yeongjae Choi, Channoh Kim, et al. 2021. “Sparsity-Aware and Re-Configurable NPU Architecture for Samsung Flagship Mobile SoC.” 2021 ACM/IEEE 48th Annual International Symposium on Computer Architecture (ISCA), June. https://doi.org/10.1109/isca52012.2021.00011.
Jang, Won, Dan Lim, Jaesam Yoon, Bongwan Kim, and Juntae Kim. 2021. “UnivNet: A Neural Vocoder with Multi-Resolution Spectrogram Discriminators for High-Fidelity Waveform Generation.” arXiv. https://doi.org/10.48550/ARXIV.2106.07889.
Janghorbani, Sepehr, and Gerard de Melo. 2023. “MultiModal Bias: Introducing a Framework for Stereotypical Bias Assessment Beyond Gender and Race in Vision Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.12734.
Jangra, Anubhav, Sourajit Mukherjee, Adam Jatowt, Sriparna Saha, and Mohammad Hasanuzzaman. 2023. “A Survey on Multi-Modal Summarization.” ACM Computing Surveys 55 (July). https://doi.org/10.1145/3584700.
Janowsky, Steven A. 1989. “Pruning Versus Clipping in Neural Networks.” Physical Review A 39 (June). https://doi.org/10.1103/physreva.39.6600.
Javadi, Golara, Kamer Ali Yuksel, Yunsu Kim, Thiago Castro Ferreira, and Mohamed Al-Badrashiny. 2024. “Word-Level ASR Quality Estimation for Efficient Corpus Sampling and Post-Editing Through Analyzing Attentions of a Reference-Free Metric.” arXiv. https://doi.org/10.48550/ARXIV.2401.11268.
Jenatton, Rodolphe, Jean-Yves Audibert, and Francis Bach. 2009. “Structured Variable Selection with Sparsity-Inducing Norms.” arXiv. https://doi.org/10.48550/ARXIV.0904.3523.
Jeni, Laszlo A., Jeffrey F. Cohn, and Fernando De La Torre. 2013. “Facing Imbalanced Data–Recommendations for the Use of Performance Metrics.” 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, September. https://doi.org/10.1109/acii.2013.47.
Jeong, Heeryeol, Taeyong Park, Seungwoo Khang, Kyoyeong Koo, Juneseuk Shin, Kyung Won Kim, and Jeongjin Lee. 2022. “Non-Rigid Registration Based on Hierarchical Deformation of Coronary Arteries in CCTA Images.” Biomedical Engineering Letters 13 (December). https://doi.org/10.1007/s13534-022-00254-8.
Jeong, Myeonghun, Hyeongju Kim, Sung Jun Cheon, Byoung Jin Choi, and Nam Soo Kim. 2021. “Diff-TTS: A Denoising Diffusion Model for Text-to-Speech.” Interspeech 2021, August. https://doi.org/10.21437/interspeech.2021-469.
Jerfel, Ghassen, Erin Grant, Thomas L. Griffiths, and Katherine Heller. 2018. “Reconciling Meta-Learning and Continual Learning with Online Mixtures of Tasks.” arXiv. https://doi.org/10.48550/ARXIV.1812.06080.
Jessen, Urszula, Michal Sroka, and Dirk Fahland. 2023. “Chit-Chat or Deep Talk: Prompt Engineering for Process Mining.” arXiv. https://doi.org/10.48550/ARXIV.2307.09909.
Ji, Bo, and Tianyi Chen. 2019. “Generative Adversarial Network for Handwritten Text.” arXiv. https://doi.org/10.48550/ARXIV.1907.11845.
Ji, Shiyu, Jinjin Shao, Daniel Agun, and Tao Yang. 2018. “Privacy-Aware Ranking with Tree Ensembles on the Cloud.” The 41st International ACM SIGIR Conference on Research &Amp; Development in Information Retrieval, June. https://doi.org/10.1145/3209978.3210022.
Ji, Yugang, Guanyi Chu, Xiao Wang, Chuan Shi, Jianan Zhao, and Junping Du. 2022. “Prohibited Item Detection via Risk Graph Structure Learning.” Proceedings of the ACM Web Conference 2022, April. https://doi.org/10.1145/3485447.3512190.
Ji, Yugang, Chuan Shi, and Xiao Wang. 2021. “Prohibited Item Detection on Heterogeneous Risk Graphs.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3481945.
Ji, Ziwei, Nayeon Lee, Rita Frieske, Tiezheng Yu, Dan Su, Yan Xu, Etsuko Ishii, Ye Jin Bang, Andrea Madotto, and Pascale Fung. 2023. “Survey of Hallucination in Natural Language Generation.” ACM Computing Surveys 55 (March). https://doi.org/10.1145/3571730.
Jia, Haipeng, Xueshuang Xiang, Da Fan, Meiyu Huang, Changhao Sun, and Yang He. 2018. “Stochastic Model Pruning via Weight Dropping Away and Back.” arXiv. https://doi.org/10.48550/ARXIV.1812.02035.
Jia, Jinyuan, Binghui Wang, Xiaoyu Cao, and Neil Zhenqiang Gong. 2020. “Certified Robustness of Community Detection Against Adversarial Structural Perturbation via Randomized Smoothing.” Proceedings of The Web Conference 2020, April. https://doi.org/10.1145/3366423.3380029.
Jia, Kui. 2016. “Improving Training of Deep Neural Networks via Singular Value Bounding.” arXiv. https://doi.org/10.48550/ARXIV.1611.06013.
Jia, Ruoxi, David Dao, Boxin Wang, Frances Ann Hubis, Nezihe Merve Gurel, Bo Li, Ce Zhang, Costas Spanos, and Dawn Song. 2019. “Efficient Task-Specific Data Valuation for Nearest Neighbor Algorithms.” Proceedings of the VLDB Endowment 12 (July). https://doi.org/10.14778/3342263.3342637.
Jia, Yangqing, Evan Shelhamer, Jeff Donahue, Sergey Karayev, Jonathan Long, Ross Girshick, Sergio Guadarrama, and Trevor Darrell. 2014. “Caffe: Convolutional Architecture for Fast Feature Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1408.5093.
Jia, Yichen, and Jong-Hyeon Jeong. 2020. “Deep Learning for Quantile Regression Under Right Censoring: DeepQuantreg.” arXiv. https://doi.org/10.48550/ARXIV.2007.07056.
Jia, Zeyang, Eli Ben-Michael, and Kosuke Imai. 2023. “Bayesian Safe Policy Learning with Chance Constrained Optimization: Application to Military Security Assessment During the Vietnam War.” arXiv. https://doi.org/10.48550/ARXIV.2307.08840.
Jiang, Jingxing, Zhubin Wang, Fei Fang, and Binqiang Zhao. 2020. “TPG-DNN: A Method for User Intent Prediction Based on Total Probability Formula and GRU Loss with Multi-Task Learning.” arXiv. https://doi.org/10.48550/ARXIV.2008.02122.
Jiang, Liang, Peter C. B. Phillips, Yubo Tao, and Yichong Zhang. 2021. “Regression-Adjusted Estimation of Quantile Treatment Effects Under Covariate-Adaptive Randomizations.” arXiv. https://doi.org/10.48550/ARXIV.2105.14752.
Jiang, Lu, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. 2017. “MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels.” arXiv. https://doi.org/10.48550/ARXIV.1712.05055.
Jiang, Minhao, Ken Ziyu Liu, Ming Zhong, Rylan Schaeffer, Siru Ouyang, Jiawei Han, and Sanmi Koyejo. 2024. “Investigating Data Contamination for Pre-Training Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.06059.
Jiang, Qing-Yuan, and Wu-Jun Li. 2016. “Deep Cross-Modal Hashing.” arXiv. https://doi.org/10.48550/ARXIV.1602.02255.
Jiang, Yuang, Shiqiang Wang, Victor Valls, Bong Jun Ko, Wei-Han Lee, Kin K. Leung, and Leandros Tassiulas. 2019. “Model Pruning Enables Efficient Federated Learning on Edge Devices.” arXiv. https://doi.org/10.48550/ARXIV.1909.12326.
Jiang, Zhengbao, Frank F. Xu, Luyu Gao, Zhiqing Sun, Qian Liu, Jane Dwivedi-Yu, Yiming Yang, Jamie Callan, and Graham Neubig. 2023. “Active Retrieval Augmented Generation.” arXiv. https://doi.org/10.48550/ARXIV.2305.06983.
Jiang, Zhenyu, Yifeng Zhu, Maxwell Svetlik, Kuan Fang, and Yuke Zhu. 2021. “Synergies Between Affordance and Geometry: 6-DoF Grasp Detection via Implicit Representations.” arXiv. https://doi.org/10.48550/ARXIV.2104.01542.
Jiao, Xiaoqi, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, Fang Wang, and Qun Liu. 2019. “TinyBERT: Distilling BERT for Natural Language Understanding.” arXiv. https://doi.org/10.48550/ARXIV.1909.10351.
Jiao, Yuling, Yanming Lai, Yisu Lo, Yang Wang, and Yunfei Yang. 2021. “Error Analysis of Deep Ritz Methods for Elliptic Equations.” arXiv. https://doi.org/10.48550/ARXIV.2107.14478.
Jin, Di, Ryan A. Rossi, Eunyee Koh, Sungchul Kim, Anup Rao, and Danai Koutra. 2019. “Latent Network Summarization.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330992.
Jin, Feihu, Jinliang Lu, Jiajun Zhang, and Chengqing Zong. 2022. “Instance-Aware Prompt Learning for Language Understanding and Generation.” arXiv. https://doi.org/10.48550/ARXIV.2201.07126.
Jin, Jonghoon, Aysegul Dundar, and Eugenio Culurciello. 2014. “Flattened Convolutional Neural Networks for Feedforward Acceleration.” arXiv. https://doi.org/10.48550/ARXIV.1412.5474.
Jin, Long, Justin Lazarow, and Zhuowen Tu. 2017. “Introspective Classification with Convolutional Nets.” arXiv. https://doi.org/10.48550/ARXIV.1704.07816.
Jin, Ming, Yixin Liu, Yu Zheng, Lianhua Chi, Yuan-Fang Li, and Shirui Pan. 2021. “ANEMONE.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3482057.
Jin, Sian, Sheng Di, Xin Liang, Jiannan Tian, Dingwen Tao, and Franck Cappello. 2019. “DeepSZ.” Proceedings of the 28th International Symposium on High-Performance Parallel and Distributed Computing, June. https://doi.org/10.1145/3307681.3326608.
Jing, Longlong, and Yingli Tian. 2019. “Self-Supervised Visual Feature Learning with Deep Neural Networks: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.1902.06162.
Jing, Luyang, Taiyong Wang, Ming Zhao, and Peng Wang. 2017. “An Adaptive Multi-Sensor Data Fusion Method Based on Deep Convolutional Neural Networks for Fault Diagnosis of Planetary Gearbox.” Sensors 17 (February). https://doi.org/10.3390/s17020414.
Jirayucharoensak, Suwicha, Setha Pan-Ngum, and Pasin Israsena. 2014. “EEG-Based Emotion Recognition Using Deep Learning Network with Principal Component Based Covariate Shift Adaptation.” The Scientific World Journal 2014. https://doi.org/10.1155/2014/627892.
Johansson, Fredrik D., Uri Shalit, and David Sontag. 2016. “Learning Representations for Counterfactual Inference.” arXiv. https://doi.org/10.48550/ARXIV.1605.03661.
Johansson, Kurt. 2002. “Non-Intersecting Paths, Random Tilings and Random Matrices.” Probability Theory and Related Fields 123 (June). https://doi.org/10.1007/s004400100187.
John, George H., and Pat Langley. 2013. “Estimating Continuous Distributions in Bayesian Classifiers.” arXiv. https://doi.org/10.48550/ARXIV.1302.4964.
Johnander, Joakim, Johan Edstedt, Michael Felsberg, Fahad Shahbaz Khan, and Martin Danelljan. 2021. “Dense Gaussian Processes for Few-Shot Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.2110.03674.
Johndrow, James E., and Kristian Lum. 2019. “An Algorithm for Removing Sensitive Information: Application to Race-Independent Recidivism Prediction.” The Annals of Applied Statistics 13 (March). https://doi.org/10.1214/18-aoas1201.
Johnson, Jeff, Matthijs Douze, and Hervé Jégou. 2017. “Billion-Scale Similarity Search with GPUs.” arXiv. https://doi.org/10.48550/ARXIV.1702.08734.
Johnson, Leah R., Robert B. Gramacy, Jeremy Cohen, Erin Mordecai, Courtney Murdock, Jason Rohr, Sadie J. Ryan, Anna M. Stewart-Ibarra, and Daniel Weikel. 2018. “Phenomenological Forecasting of Disease Incidence Using Heteroskedastic Gaussian Processes: A Dengue Case Study.” The Annals of Applied Statistics 12 (March). https://doi.org/10.1214/17-aoas1090.
Johnson, Melvin, Mike Schuster, Quoc V. Le, Maxim Krikun, Yonghui Wu, Zhifeng Chen, Nikhil Thorat, et al. 2016. “Google’s Multilingual Neural Machine Translation System: Enabling Zero-Shot Translation.” arXiv. https://doi.org/10.48550/ARXIV.1611.04558.
Johnson, Rie, and Tong Zhang. 2014. “Effective Use of Word Order for Text Categorization with Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1412.1058.
Jombart, Thibaut, Stéphane Ghozzi, Dirk Schumacher, Quentin J Leclerc, Mark Jit, Stefan Flasche, Felix Greaves, et al. 2020. “Real-Time Monitoring of COVID-19 Dynamics Using Automated Trend Fitting and Anomaly Detection,” September. https://doi.org/10.1101/2020.09.02.20186502.
Jones, Alex, William Yang Wang, and Kyle Mahowald. 2021. “A Massively Multilingual Analysis of Cross-Linguality in Shared Embedding Space.” arXiv. https://doi.org/10.48550/ARXIV.2109.06324.
Jordan, Kareem L., and Tina L. Freiburger. 2015. “The Effect of Race/Ethnicity on Sentencing: Examining Sentence Type, Jail Length, and Prison Length.” Journal of Ethnicity in Criminal Justice 13 (January). https://doi.org/10.1080/15377938.2014.984045.
Jordao, Artur, Ricardo Kloss, Fernando Yamada, and William Robson Schwartz. 2018. “Pruning Deep Neural Networks Using Partial Least Squares.” arXiv. https://doi.org/10.48550/ARXIV.1810.07610.
Joshi, Bhautik, Kristen Stewart, and David Shapiro. 2017. “Bringing Impressionism to Life with Neural Style Transfer in Come Swim.” arXiv. https://doi.org/10.48550/ARXIV.1701.04928.
Joshi, Mandar, Eunsol Choi, Daniel Weld, and Luke Zettlemoyer. 2017. “TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension.” Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/p17-1147.
Josifoski, Martin, Marija Sakota, Maxime Peyrard, and Robert West. 2023. “Exploiting Asymmetry for Synthetic Training Data Generation: SynthIE and the Case of Information Extraction.” arXiv. https://doi.org/10.48550/ARXIV.2303.04132.
Joulin, Armand, Edouard Grave, Piotr Bojanowski, Matthijs Douze, Hérve Jégou, and Tomas Mikolov. 2016. “FastText.zip: Compressing Text Classification Models.” arXiv. https://doi.org/10.48550/ARXIV.1612.03651.
Joulin, Armand, Edouard Grave, Piotr Bojanowski, and Tomas Mikolov. 2016. “Bag of Tricks for Efficient Text Classification.” arXiv. https://doi.org/10.48550/ARXIV.1607.01759.
Jouppi, Norm, George Kurian, Sheng Li, Peter Ma, Rahul Nagarajan, Lifeng Nai, Nishant Patil, et al. 2023. “TPU V4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings.” Proceedings of the 50th Annual International Symposium on Computer Architecture, June. https://doi.org/10.1145/3579371.3589350.
Jozefowicz, Rafal, Oriol Vinyals, Mike Schuster, Noam Shazeer, and Yonghui Wu. 2016. “Exploring the Limits of Language Modeling.” arXiv. https://doi.org/10.48550/ARXIV.1602.02410.
Ju, Cheng, Aurélien Bibaut, and Mark J. van der Laan. 2017. “The Relative Performance of Ensemble Methods with Deep Convolutional Neural Networks for Image Classification.” arXiv. https://doi.org/10.48550/ARXIV.1704.01664.
Juan, Yuchin, Damien Lefortier, and Olivier Chapelle. 2017. “Field-Aware Factorization Machines in a Real-World Online Advertising System.” Proceedings of the 26th International Conference on World Wide Web Companion - WWW ’17 Companion. https://doi.org/10.1145/3041021.3054185.
Juefei-Xu, Felix, Vishnu Naresh Boddeti, and Marios Savvides. 2016. “Local Binary Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1608.06049.
Jumper, John, Richard Evans, Alexander Pritzel, Tim Green, Michael Figurnov, Olaf Ronneberger, Kathryn Tunyasuvunakool, et al. 2021. “Highly Accurate Protein Structure Prediction with AlphaFold.” Nature 596 (July). https://doi.org/10.1038/s41586-021-03819-2.
Kaddour, Jean, Joshua Harris, Maximilian Mozes, Herbie Bradley, Roberta Raileanu, and Robert McHardy. 2023. “Challenges and Applications of Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2307.10169.
Kadlec, Rudolf, Ondrej Bajgar, and Jan Kleindienst. 2017. “Knowledge Base Completion: Baselines Strike Back.” arXiv. https://doi.org/10.48550/ARXIV.1705.10744.
Kadlec, Rudolf, Martin Schmid, Ondrej Bajgar, and Jan Kleindienst. 2016. “Text Understanding with the Attention Sum Reader Network.” arXiv. https://doi.org/10.48550/ARXIV.1603.01547.
Kaelbling, L. P., M. L. Littman, and A. W. Moore. 1996. “Reinforcement Learning: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.CS/9605103.
Kaghazgaran, Parisa, James Caverlee, and Anna Squicciarini. 2018. “Combating Crowdsourced Review Manipulators.” Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/3159652.3159726.
Kairouz, Peter, H. Brendan McMahan, Brendan Avent, Aurélien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Kallista Bonawitz, et al. 2019. “Advances and Open Problems in Federated Learning.” arXiv. https://doi.org/10.48550/ARXIV.1912.04977.
Kalchbrenner, Nal, Edward Grefenstette, and Phil Blunsom. 2014. “A Convolutional Neural Network for Modelling Sentences.” arXiv. https://doi.org/10.48550/ARXIV.1404.2188.
Kallus, Nathan, Xiaojie Mao, and Masatoshi Uehara. 2021. “Causal Inference Under Unmeasured Confounding with Negative Controls: A Minimax Learning Approach.” arXiv. https://doi.org/10.48550/ARXIV.2103.14029.
Kalyanakrishnan, Shivaram, Deepthi Singh, and Ravi Kant. 2014. “On Building Decision Trees from Large-Scale Data in Applications of on-Line Advertising.” Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/2661829.2662044.
Kameoka, Hirokazu, Takuhiro Kaneko, Kou Tanaka, and Nobukatsu Hojo. 2019. “ACVAE-VC: Non-Parallel Voice Conversion with Auxiliary Classifier Variational Autoencoder.” IEEE/ACM Transactions on Audio, Speech, and Language Processing 27 (September). https://doi.org/10.1109/taslp.2019.2917232.
Kaminski, Michael M., Omar O. Abudayyeh, Jonathan S. Gootenberg, Feng Zhang, and James J. Collins. 2021. “CRISPR-Based Diagnostics.” Nature Biomedical Engineering 5 (July). https://doi.org/10.1038/s41551-021-00760-7.
Kamiran, Faisal, Toon Calders, and Mykola Pechenizkiy. 2010. “Discrimination Aware Decision Tree Learning.” 2010 IEEE International Conference on Data Mining, December. https://doi.org/10.1109/icdm.2010.50.
Kamlish, Isaac, Isaac Bentata Chocron, and Nicholas McCarthy. 2019. “SentiMATE: Learning to Play Chess Through Natural Language Processing.” arXiv. https://doi.org/10.48550/ARXIV.1907.08321.
Kammoun, Amina, Rim Slama, Hedi Tabia, Tarek Ouni, and Mohmed Abid. 2022. “Generative Adversarial Networks for Face Generation: A Survey.” ACM Computing Surveys, March. https://doi.org/10.1145/1122445.1122456.
Kamnitsas, Konstantinos, Christian Baumgartner, Christian Ledig, Virginia F. J. Newcombe, Joanna P. Simpson, Andrew D. Kane, David K. Menon, et al. 2016. “Unsupervised Domain Adaptation in Brain Lesion Segmentation with Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1612.08894.
Kanagal, B., A. Ahmed, S. Pandey, V. Josifovski, L. Garcia-Pueyo, and J. Yuan. 2013. “Focused Matrix Factorization for Audience Selection in Display Advertising.” 2013 IEEE 29th International Conference on Data Engineering (ICDE), April. https://doi.org/10.1109/icde.2013.6544841.
Kanazawa, Angjoo, Abhishek Sharma, and David Jacobs. 2014. “Locally Scale-Invariant Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1412.5104.
Kang, Daniel, John Emmons, Firas Abuzaid, Peter Bailis, and Matei Zaharia. 2017. “NoScope: Optimizing Neural Network Queries over Video at Scale.” arXiv. https://doi.org/10.48550/ARXIV.1703.02529.
Kang, Daniel, Xuechen Li, Ion Stoica, Carlos Guestrin, Matei Zaharia, and Tatsunori Hashimoto. 2023. “Exploiting Programmatic Behavior of LLMs: Dual-Use Through Standard Security Attacks.” arXiv. https://doi.org/10.48550/ARXIV.2302.05733.
Kang, Daniel, Yi Sun, Tom Brown, Dan Hendrycks, and Jacob Steinhardt. 2019. “Transfer of Adversarial Robustness Between Perturbation Types.” arXiv. https://doi.org/10.48550/ARXIV.1905.01034.
Kang, Jingqi, Tongtong Wu, Jinming Zhao, Guitao Wang, Guilin Qi, Yuan-Fang Li, and Gholamreza Haffari. 2024. “Towards Event Extraction from Speech with Contextual Clues.” arXiv. https://doi.org/10.48550/ARXIV.2401.15385.
Kang, Lei, Pau Riba, Marcal Rusinol, Alicia Fornes, and Mauricio Villegas. 2022. “Content and Style Aware Generation of Text-Line Images for Handwriting Recognition.” IEEE Transactions on Pattern Analysis and Machine Intelligence 44 (December). https://doi.org/10.1109/tpami.2021.3122572.
Kang, Lei, Pau Riba, Marçal Rusiñol, Alicia Fornés, and Mauricio Villegas. 2020. “Pay Attention to What You Read: Non-Recurrent Handwritten Text-Line Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2005.13044.
Kano, Yoshinobu, Mi-Young Kim, Masaharu Yoshioka, Yao Lu, Juliano Rabelo, Naoki Kiyota, Randy Goebel, and Ken Satoh. 2019. “COLIEE-2018: Evaluation of the Competition on Legal Information Extraction and Entailment.” New Frontiers in Artificial Intelligence. https://doi.org/10.1007/978-3-030-31605-1_14.
Kaplan, Russell, Christopher Sauer, and Alexander Sosa. 2017. “Beating Atari with Natural Language Guided Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1704.05539.
Karim, Md. Rezaul, Felix Hermsen, Sisay Adugna Chala, Paola de Perthuis, and Avikarsha Mandal. 2023. “Catch Me If You Can: Semi-Supervised Graph Learning for Spotting Money Laundering.” arXiv. https://doi.org/10.48550/ARXIV.2302.11880.
Karimi, Akbar, Leonardo Rossi, and Andrea Prati. 2020. “Adversarial Training for Aspect-Based Sentiment Analysis with BERT.” arXiv. https://doi.org/10.48550/ARXIV.2001.11316.
Karlsson, Niklas. 2020. “Feedback Control in Programmatic Advertising: The Frontier of Optimization in Real-Time Bidding.” IEEE Control Systems 40 (October). https://doi.org/10.1109/mcs.2020.3005013.
Karmakar, Sayar, and Arkaprava Roy. 2021. “Bayesian Modelling of Time-Varying Conditional Heteroscedasticity.” Bayesian Analysis 16 (December). https://doi.org/10.1214/21-ba1267.
Karpathy, Andrej, and Li Fei-Fei. 2017. “Deep Visual-Semantic Alignments for Generating Image Descriptions.” IEEE Transactions on Pattern Analysis and Machine Intelligence 39 (April). https://doi.org/10.1109/tpami.2016.2598339.
Karpathy, Andrej, Justin Johnson, and Li Fei-Fei. 2015. “Visualizing and Understanding Recurrent Networks.” arXiv. https://doi.org/10.48550/ARXIV.1506.02078.
Karras, Tero, Samuli Laine, and Timo Aila. 2018. “A Style-Based Generator Architecture for Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1812.04948.
———. 2019. “A Style-Based Generator Architecture for Generative Adversarial Networks.” 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr.2019.00453.
Karwath, Andreas, Kristian Kersting, and Niels Landwehr. 2008. “Boosting Relational Sequence Alignments.” 2008 Eighth IEEE International Conference on Data Mining, December. https://doi.org/10.1109/icdm.2008.127.
Kasanishi, Tetsu, Xueting Wang, and Toshihiko Yamasaki. 2021. “Edge-Level Explanations for Graph Neural Networks by Extending Explainability Methods for Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2111.00722.
Kashefi, Ali, and Tapan Mukerji. 2023. “ChatGPT for Programming Numerical Methods.” arXiv. https://doi.org/10.48550/ARXIV.2303.12093.
Kass, Robert E., Bradley P. Carlin, Andrew Gelman, and Radford M. Neal. 1998. “Markov Chain Monte Carlo in Practice: A Roundtable Discussion.” The American Statistician 52 (May). https://doi.org/10.1080/00031305.1998.10480547.
Kataria, Saurabh, Phani Sankar Nidadavolu, Jesús Villalba, Nanxin Chen, Paola García, and Najim Dehak. 2019. “Feature Enhancement with Deep Feature Losses for Speaker Verification.” arXiv. https://doi.org/10.48550/ARXIV.1910.11905.
Katevas, Kleomenis, Ilias Leontiadis, Martin Pielot, and Joan Serrà. 2017. “Continual Prediction of Notification Attendance with Classical and Deep Network Approaches.” arXiv. https://doi.org/10.48550/ARXIV.1712.07120.
Katt, Sammie, Frans Oliehoek, and Christopher Amato. 2018. “Bayesian Reinforcement Learning in Factored POMDPs.” arXiv. https://doi.org/10.48550/ARXIV.1811.05612.
Katz, Daniel Martin, Dirk Hartung, Lauritz Gerlach, Abhik Jana, and Michael J. Bommarito. 2023. “Natural Language Processing in the Legal Domain.” arXiv. https://doi.org/10.48550/ARXIV.2302.12039.
Katz, Guy, Clark Barrett, David Dill, Kyle Julian, and Mykel Kochenderfer. 2017. “Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1702.01135.
Kaul, Chaitanya, Joshua Mitton, Hang Dai, and Roderick Murray-Smith. 2021. “CpT: Convolutional Point Transformer for 3D Point Cloud Processing.” arXiv. https://doi.org/10.48550/ARXIV.2111.10866.
Kawaguchi, K., Y. Bengio, and L. Kaelbling. 2022. “Generalization in Deep Learning.” Mathematical Aspects of Deep Learning, December. https://doi.org/10.1017/9781009025096.003.
Kawaguchi, Kenji. 2016. “Deep Learning Without Poor Local Minima.” arXiv. https://doi.org/10.48550/ARXIV.1605.07110.
Kawahara, Yoshinobu, Kenneth Bollen, Shohei Shimizu, and Takashi Washio. 2010. “GroupLiNGAM: Linear Non-Gaussian Acyclic Models for Sets of Variables.” arXiv. https://doi.org/10.48550/ARXIV.1006.5041.
Kazemi, Seyed Mehran, Rishab Goel, Kshitij Jain, Ivan Kobyzev, Akshay Sethi, Peter Forsyth, and Pascal Poupart. 2019. “Representation Learning for Dynamic Graphs: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.1905.11485.
Kegelmeyer, W. Philip, Jeremy D Wendt, and Ali Pinar. 2018. “An Example of Counter-Adversarial Community Detection Analysis,” October. https://doi.org/10.2172/1481570.
Kégl, Balázs. 2013. “The Return of AdaBoost.MH: Multi-Class Hamming Trees.” arXiv. https://doi.org/10.48550/ARXIV.1312.6086.
Keith, Katherine A., Abram Handler, Michael Pinkham, Cara Magliozzi, Joshua McDuffie, and Brendan O’Connor. 2017. “Identifying Civilians Killed by Police with Distantly Supervised Entity-Event Extraction.” arXiv. https://doi.org/10.48550/ARXIV.1707.07086.
Kelkar, Amol, Rohan Relan, Vaishali Bhardwaj, Saurabh Vaichal, Chandra Khatri, and Peter Relan. 2020. “Bertrand-DR: Improving Text-to-SQL Using a Discriminative Re-Ranker,” February. http://arxiv.org/abs/2002.00557v2.
Kenyon, Richard. 2000. “The Asymptotic Determinant of the Discrete Laplacian.” arXiv. https://doi.org/10.48550/ARXIV.MATH-PH/0011042.
Kerschke, Pascal, Holger H. Hoos, Frank Neumann, and Heike Trautmann. 2018. “Automated Algorithm Selection: Survey and Perspectives.” arXiv. https://doi.org/10.48550/ARXIV.1811.11597.
Keuper, Margret, Siyu Tang, Yu Zhongjie, Bjoern Andres, Thomas Brox, and Bernt Schiele. 2016. “A Multi-Cut Formulation for Joint Segmentation and Tracking of Multiple Objects.” arXiv. https://doi.org/10.48550/ARXIV.1607.06317.
Keyvanrad, Mohammad Ali, and Mohammad Mehdi Homayounpour. 2014. “A Brief Survey on Deep Belief Networks and Introducing a New Object Oriented Toolbox (DeeBNet).” arXiv. https://doi.org/10.48550/ARXIV.1408.3264.
Khachatryan, Levon, Andranik Movsisyan, Vahram Tadevosyan, Roberto Henschel, Zhangyang Wang, Shant Navasardyan, and Humphrey Shi. 2023. “Text2Video-Zero: Text-to-Image Diffusion Models Are Zero-Shot Video Generators.” arXiv. https://doi.org/10.48550/ARXIV.2303.13439.
Khalil, Mohammad, and Erkan Er. 2023. “Will ChatGPT Get You Caught? Rethinking of Plagiarism Detection.” arXiv. https://doi.org/10.48550/ARXIV.2302.04335.
Khan, Latif U., Walid Saad, Zhu Han, Ekram Hossain, and Choong Seon Hong. 2020. “Federated Learning for Internet of Things: Recent Advances, Taxonomy, and Open Challenges.” arXiv. https://doi.org/10.48550/ARXIV.2009.13012.
Khan, Salman, Muzammal Naseer, Munawar Hayat, Syed Waqas Zamir, Fahad Shahbaz Khan, and Mubarak Shah. 2022. “Transformers in Vision: A Survey.” ACM Computing Surveys 54 (January). https://doi.org/10.1145/3505244.
Khanday, Akib Mohi Ud Din, Syed Tanzeel Rabani, Qamar Rayees Khan, Nusrat Rouf, and Masarat Mohi Ud Din. 2020. “Machine Learning Based Approaches for Detecting COVID-19 Using Clinical Text Data.” International Journal of Information Technology 12 (June). https://doi.org/10.1007/s41870-020-00495-9.
Khattab, Omar, and Matei Zaharia. 2020. “ColBERT.” Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3397271.3401075.
Khattak, Muhammad Uzair, Muhammad Ferjad Naeem, Muzammal Naseer, Luc Van Gool, and Federico Tombari. 2024. “Learning to Prompt with Text Only Supervision for Vision-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.02418.
Khattak, Muhammad Uzair, Hanoona Rasheed, Muhammad Maaz, Salman Khan, and Fahad Shahbaz Khan. 2022. “MaPLe: Multi-Modal Prompt Learning.” arXiv. https://doi.org/10.48550/ARXIV.2210.03117.
Khayati, Mourad, Ines Arous, Zakhar Tymchenko, and Philippe Cudré-Mauroux. 2020. “ORBITS.” Proceedings of the VLDB Endowment 14 (November). https://doi.org/10.14778/3430915.3430920.
Khayati, Mourad, Alberto Lerner, Zakhar Tymchenko, and Philippe Cudré-Mauroux. 2020. “Mind the Gap.” Proceedings of the VLDB Endowment 13 (January). https://doi.org/10.14778/3377369.3377383.
Khodak, Mikhail, Nikunj Saunshi, and Kiran Vodrahalli. 2017. “A Large Self-Annotated Corpus for Sarcasm.” arXiv. https://doi.org/10.48550/ARXIV.1704.05579.
Khoreva, Anna, Rodrigo Benenson, Eddy Ilg, Thomas Brox, and Bernt Schiele. 2017. “Lucid Data Dreaming for Video Object Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1703.09554.
Khurana, Udayan, Horst Samulowitz, and Deepak Turaga. 2017. “Feature Engineering for Predictive Modeling Using Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1709.07150.
Kilner, James M., and Karl J. Friston. 2010. “Topological Inference for EEG and MEG.” The Annals of Applied Statistics 4 (September). https://doi.org/10.1214/10-aoas337.
Kim, Do-Guk, and Heung-Chang Lee. 2020. “Differentiable Neural Architecture Transformation for Reproducible Architecture Improvement.” arXiv. https://doi.org/10.48550/ARXIV.2006.08231.
Kim, Edward, Kevin Huang, Alex Tomala, Sara Matthews, Emma Strubell, Adam Saunders, Andrew McCallum, and Elsa Olivetti. 2017. “Machine-Learned and Codified Synthesis Parameters of Oxide Materials.” Scientific Data 4 (September). https://doi.org/10.1038/sdata.2017.127.
Kim, Heehyeon, Jinhyeok Choi, and Joyce Jiyoung Whang. 2023. “Dynamic Relation-Attentive Graph Neural Networks for Fraud Detection.” arXiv. https://doi.org/10.48550/ARXIV.2310.04171.
Kim, Hyeongwoo, Michael Zollhöfer, Ayush Tewari, Justus Thies, Christian Richardt, and Christian Theobalt. 2017. “InverseFaceNet: Deep Monocular Inverse Face Rendering.” arXiv. https://doi.org/10.48550/ARXIV.1703.10956.
Kim, Hyesung, Jihong Park, Mehdi Bennis, and Seong-Lyun Kim. 2018. “Blockchained on-Device Federated Learning.” arXiv. https://doi.org/10.48550/ARXIV.1808.03949.
Kim, Hyunjik, and Andriy Mnih. 2018. “Disentangling by Factorising.” arXiv. https://doi.org/10.48550/ARXIV.1802.05983.
Kim, Jaeyoung, Mostafa El-Khamy, and Jungwon Lee. 2017. “Residual LSTM: Design of a Deep Recurrent Architecture for Distant Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1701.03360.
Kim, Jangho, Minsung Hyun, Inseop Chung, and Nojun Kwak. 2019. “Feature Fusion for Online Mutual Knowledge Distillation.” arXiv. https://doi.org/10.48550/ARXIV.1904.09058.
Kim, Jangho, SeongUk Park, and Nojun Kwak. 2018. “Paraphrasing Complex Network: Network Compression via Factor Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1802.04977.
Kim, Jin Hee, Brett Grady, Ruolong Lian, John Brothers, and Jason H. Anderson. 2017. “FPGA-Based CNN Inference Accelerator Synthesized from Multi-Threaded c Software.” 2017 30th IEEE International System-on-Chip Conference (SOCC), September. https://doi.org/10.1109/socc.2017.8226056.
Kim, Jiseong, and Key-sun Choi. 2020. “Unsupervised Fact Checking by Counter-Weighted Positive and Negative Evidential Paths in a Knowledge Graph.” Proceedings of the 28th International Conference on Computational Linguistics. https://doi.org/10.18653/v1/2020.coling-main.147.
Kim, Joyce, Stefano Schiavon, and Gail Brager. 2018. “Personal Comfort Models – a New Paradigm in Thermal Comfort for Occupant-Centric Environmental Control.” Building and Environment 132 (March). https://doi.org/10.1016/j.buildenv.2018.01.023.
Kim, Kwangyoun, Felix Wu, Yifan Peng, Jing Pan, Prashant Sridhar, Kyu J. Han, and Shinji Watanabe. 2022. “E-Branchformer: Branchformer with Enhanced Merging for Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2210.00077.
Kim, Minje, and Paris Smaragdis. 2016. “Bitwise Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1601.06071.
Kim, Minsu, Hyung-Il Kim, and Yong Man Ro. 2023. “Prompt Tuning of Deep Neural Networks for Speaker-Adaptive Visual Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2302.08102.
Kim, Minyoung. 2022. “Gaussian Process Modeling of Approximate Inference Errors for Variational Autoencoders.” 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52688.2022.00034.
Kim, Minyoung, Pritish Sahu, Behnam Gholami, and Vladimir Pavlovic. 2019. “Unsupervised Visual Domain Adaptation: A Deep Max-Margin Gaussian Process Approach.” arXiv. https://doi.org/10.48550/ARXIV.1902.08727.
Kim, Sehoon, Coleman Hooper, Amir Gholami, Zhen Dong, Xiuyu Li, Sheng Shen, Michael W. Mahoney, and Kurt Keutzer. 2023. “SqueezeLLM: Dense-and-Sparse Quantization.” arXiv. https://doi.org/10.48550/ARXIV.2306.07629.
Kim, Seung-Jean, Kwangmoo Koh, Stephen Boyd, and Dimitry Gorinevsky. 2009. \(\ell_1\) Trend Filtering.” SIAM Review 51 (May). https://doi.org/10.1137/070690274.
Kim, Sundong, Yu-Che Tsai, Karandeep Singh, Yeonsoo Choi, Etim Ibok, Cheng-Te Li, and Meeyoung Cha. 2020. “DATE.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403339.
Kim, Sunghwan, Paul A. Thiessen, Evan E. Bolton, Jie Chen, Gang Fu, Asta Gindulyte, Lianyi Han, et al. 2015. “PubChem Substance and Compound Databases.” Nucleic Acids Research 44 (September). https://doi.org/10.1093/nar/gkv951.
Kim, Suyoun, Takaaki Hori, and Shinji Watanabe. 2016. “Joint CTC-Attention Based End-to-End Speech Recognition Using Multi-Task Learning.” arXiv. https://doi.org/10.48550/ARXIV.1609.06773.
Kim, Suyoun, Yuan Shangguan, Jay Mahadeokar, Antoine Bruguier, Christian Fuegen, Michael L. Seltzer, and Duc Le. 2020. “Improved Neural Language Model Fusion for Streaming Recurrent Neural Network Transducer.” arXiv. https://doi.org/10.48550/ARXIV.2010.13878.
Kim, Taeksoo, Moonsu Cha, Hyunsoo Kim, Jung Kwon Lee, and Jiwon Kim. 2017. “Learning to Discover Cross-Domain Relations with Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1703.05192.
Kim, Yong-Deok, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu Yang, and Dongjun Shin. 2015. “Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications.” arXiv. https://doi.org/10.48550/ARXIV.1511.06530.
Kim, Yoon. 2014. “Convolutional Neural Networks for Sentence Classification.” arXiv. https://doi.org/10.48550/ARXIV.1408.5882.
Kim, Yoon, Yacine Jernite, David Sontag, and Alexander M. Rush. 2015. “Character-Aware Neural Language Models.” arXiv. https://doi.org/10.48550/ARXIV.1508.06615.
Kim, Yoon, and Alexander M. Rush. 2016. “Sequence-Level Knowledge Distillation.” arXiv. https://doi.org/10.48550/ARXIV.1606.07947.
Kim, Yulhwa, Hyungjun Kim, and Jae-Joon Kim. 2018. “Neural Network-Hardware Co-Design for Scalable RRAM-Based BNN Accelerators.” arXiv. https://doi.org/10.48550/ARXIV.1811.02187.
King, Brian, and Daniel R. Kowal. 2023. “Warped Dynamic Linear Models for Time Series of Counts.” Bayesian Analysis -1 (January). https://doi.org/10.1214/23-ba1394.
King, Gary, Patrick Lam, and Margaret E. Roberts. 2017. “Computer‐assisted Keyword and Document Set Discovery from Unstructured Text.” American Journal of Political Science 61 (April). https://doi.org/10.1111/ajps.12291.
Kingma, Diederik P., and Jimmy Ba. 2014. “Adam: A Method for Stochastic Optimization,” December. http://arxiv.org/abs/1412.6980v9.
Kingma, Diederik P, and Max Welling. 2013. “Auto-Encoding Variational Bayes.” arXiv. https://doi.org/10.48550/ARXIV.1312.6114.
Kiperwasser, Eliyahu, and Yoav Goldberg. 2016. “Simple and Accurate Dependency Parsing Using Bidirectional LSTM Feature Representations.” arXiv. https://doi.org/10.48550/ARXIV.1603.04351.
Kipf, Thomas N., and Max Welling. 2016a. “Semi-Supervised Classification with Graph Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1609.02907.
———. 2016b. “Variational Graph Auto-Encoders.” arXiv. https://doi.org/10.48550/ARXIV.1611.07308.
Kirchenbauer, John, Jonas Geiping, Yuxin Wen, Jonathan Katz, Ian Miers, and Tom Goldstein. 2023. “A Watermark for Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.10226.
Kirkpatrick, James, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A. Rusu, Kieran Milan, et al. 2017. “Overcoming Catastrophic Forgetting in Neural Networks.” Proceedings of the National Academy of Sciences 114 (March). https://doi.org/10.1073/pnas.1611835114.
Kiros, Ryan, Ruslan Salakhutdinov, and Richard S. Zemel. 2014. “Unifying Visual-Semantic Embeddings with Multimodal Neural Language Models.” arXiv. https://doi.org/10.48550/ARXIV.1411.2539.
Kiros, Ryan, Yukun Zhu, Ruslan Salakhutdinov, Richard S. Zemel, Antonio Torralba, Raquel Urtasun, and Sanja Fidler. 2015. “Skip-Thought Vectors.” arXiv. https://doi.org/10.48550/ARXIV.1506.06726.
Kıcıman, Emre, Robert Ness, Amit Sharma, and Chenhao Tan. 2023. “Causal Reasoning and Large Language Models: Opening a New Frontier for Causality.” arXiv. https://doi.org/10.48550/ARXIV.2305.00050.
Klambauer, Günter, Thomas Unterthiner, Andreas Mayr, and Sepp Hochreiter. 2017. “Self-Normalizing Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1706.02515.
Klarman, Uri, Marcel Flores, and Aleksandar Kuzmanovic. 2018. “Mining the Web with Webcoin.” Proceedings of the 14th International Conference on Emerging Networking EXperiments and Technologies, December. https://doi.org/10.1145/3281411.3281415.
Klein, Benjamin, and Lior Wolf. 2017. “End-to-End Supervised Product Quantization for Image Search and Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.1711.08589.
Klein, Guillaume, Yoon Kim, Yuntian Deng, Jean Senellart, and Alexander Rush. 2017. “OpenNMT: Open-Source Toolkit for Neural Machine Translation.” Proceedings of ACL 2017, System Demonstrations. https://doi.org/10.18653/v1/p17-4012.
Kliemann, Wolfgang. 1987. “Recurrence and Invariant Measures for Degenerate Diffusions.” The Annals of Probability 15 (April). https://doi.org/10.1214/aop/1176992166.
Klokov, Roman, and Victor Lempitsky. 2017. “Escape from Cells: Deep Kd-Networks for the Recognition of 3D Point Cloud Models.” arXiv. https://doi.org/10.48550/ARXIV.1704.01222.
Knott, Alistair, Dino Pedreschi, Raja Chatila, Tapabrata Chakraborti, Susan Leavy, Ricardo Baeza-Yates, David Eyers, et al. 2023. “Generative AI Models Should Include Detection Mechanisms as a Condition for Public Release.” Ethics and Information Technology 25 (October). https://doi.org/10.1007/s10676-023-09728-4.
“Knowledge Discovery in Databases: PKDD 2003.” 2003. Lecture Notes in Computer Science. https://doi.org/10.1007/b13634.
Knowles, Antti, and Jun Yin. 2014. “Anisotropic Local Laws for Random Matrices.” arXiv. https://doi.org/10.48550/ARXIV.1410.3516.
Kocián, Matěj, Jakub Náplava, Daniel Štancl, and Vladimír Kadlec. 2021. “Siamese BERT-Based Model for Web Search Relevance Ranking Evaluated on a New Czech Dataset.” arXiv. https://doi.org/10.48550/ARXIV.2112.01810.
Koehn, Philipp, Franz Josef Och, and Daniel Marcu. 2003. “Statistical Phrase-Based Translation.” Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology - NAACL ’03. https://doi.org/10.3115/1073445.1073462.
Koh, Pang Wei, and Percy Liang. 2017. “Understanding Black-Box Predictions via Influence Functions.” arXiv. https://doi.org/10.48550/ARXIV.1703.04730.
Kokkinos, Iasonas. 2016. “UberNet: Training a ‘Universal’ Convolutional Neural Network for Low-, Mid-, and High-Level Vision Using Diverse Datasets and Limited Memory.” arXiv. https://doi.org/10.48550/ARXIV.1609.02132.
Kokoris-Kogias, Eleftherios, Philipp Jovanovic, Nicolas Gailly, Ismail Khoffi, Linus Gasser, and Bryan Ford. 2016. “Enhancing Bitcoin Security and Performance with Strong Consistency via Collective Signing.” arXiv. https://doi.org/10.48550/ARXIV.1602.06997.
Kolar, Mladen, Le Song, Amr Ahmed, and Eric P. Xing. 2010. “Estimating Time-Varying Networks.” The Annals of Applied Statistics 4 (March). https://doi.org/10.1214/09-aoas308.
Kolter, J. Zico, and Andrew Y. Ng. 2009. “Near-Bayesian Exploration in Polynomial Time.” Proceedings of the 26th Annual International Conference on Machine Learning, June. https://doi.org/10.1145/1553374.1553441.
Konečný, Jakub, H. Brendan McMahan, Felix X. Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. 2016. “Federated Learning: Strategies for Improving Communication Efficiency.” arXiv. https://doi.org/10.48550/ARXIV.1610.05492.
Kong, Jungil, Jaehyeon Kim, and Jaekyoung Bae. 2020. “HiFi-GAN: Generative Adversarial Networks for Efficient and High Fidelity Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.2010.05646.
Kong, Qiuqiang, Yin Cao, Turab Iqbal, Yuxuan Wang, Wenwu Wang, and Mark D. Plumbley. 2019. “PANNs: Large-Scale Pretrained Audio Neural Networks for Audio Pattern Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1912.10211.
Kong, Weize, Spurthi Amba Hombaiah, Mingyang Zhang, Qiaozhu Mei, and Michael Bendersky. 2024. “PRewrite: Prompt Rewriting with Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.2401.08189.
Kong, Xin, Shikun Liu, Marwan Taher, and Andrew J. Davison. 2023. “vMAP: Vectorised Object Mapping for Neural Field SLAM.” arXiv. https://doi.org/10.48550/ARXIV.2302.01838.
Konno, N., and T. Shiga. 1988. “Stochastic Partial Differential Equations for Some Measure-Valued Diffusions.” Probability Theory and Related Fields 79 (September). https://doi.org/10.1007/bf00320919.
Kononen, V. n.d. “Asymmetric Multiagent Reinforcement Learning.” IEEE/WIC International Conference on Intelligent Agent Technology, 2003. IAT 2003. https://doi.org/10.1109/iat.2003.1241094.
Koplan, Bruce A., and William G. Stevenson. 2009. “Ventricular Tachycardia and Sudden Cardiac Death.” Mayo Clinic Proceedings 84 (March). https://doi.org/10.4065/84.3.289.
Korattikara, Anoop, Vivek Rathod, Kevin Murphy, and Max Welling. 2015. “Bayesian Dark Knowledge.” arXiv. https://doi.org/10.48550/ARXIV.1506.04416.
Kos, Jernej, Ian Fischer, and Dawn Song. 2017. “Adversarial Examples for Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.1702.06832.
Kosub, Sven. 2016. “A Note on the Triangle Inequality for the Jaccard Distance.” arXiv. https://doi.org/10.48550/ARXIV.1612.02696.
Kotnis, Bhushan, and Vivi Nastase. 2017. “Analysis of the Impact of Negative Sampling on Link Prediction in Knowledge Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1708.06816.
Koupaee, Mahnaz, and William Yang Wang. 2018. “WikiHow: A Large Scale Text Summarization Dataset.” arXiv. https://doi.org/10.48550/ARXIV.1810.09305.
Koushik, Jayanth. 2016. “Understanding Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1605.09081.
Kouvaros, Panagiotis, Francesco Leofante, Blake Edwards, Calvin Chung, Dragos Margineantu, and Alessio Lomuscio. 2023. “Verification of Semantic Key Point Detection for Aircraft Pose Estimation.” Proceedings of the Twentieth International Conference on Principles of Knowledge Representation and Reasoning, September. https://doi.org/10.24963/kr.2023/77.
Kowal, Daniel R., and Antonio Canale. 2023. “Semiparametric Functional Factor Models with Bayesian Rank Selection.” Bayesian Analysis 18 (December). https://doi.org/10.1214/23-ba1410.
Kramkov, D. O. 1996. “Optional Decomposition of Supermartingales and Hedging Contingent Claims in Incomplete Security Markets.” Probability Theory and Related Fields 105 (December). https://doi.org/10.1007/bf01191909.
Kraus, Daniel, and Claudia Czado. 2017. “Growing Simplified Vine Copula Trees: Improving Dißmann’s Algorithm.” arXiv. https://doi.org/10.48550/ARXIV.1703.05203.
Kreutz, Christin Katharina, and Ralf Schenkel. 2022. “Scientific Paper Recommendation Systems: A Literature Review of Recent Publications.” International Journal on Digital Libraries 23 (October). https://doi.org/10.1007/s00799-022-00339-w.
Kreuzberger, Dominik, Niklas Kühl, and Sebastian Hirschl. 2022. “Machine Learning Operations (MLOps): Overview, Definition, and Architecture.” arXiv. https://doi.org/10.48550/ARXIV.2205.02302.
Krichene, Walid, and Steffen Rendle. 2020. “On Sampled Metrics for Item Recommendation.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403226.
Krishna, Satyapriya, Tessa Han, Alex Gu, Javin Pombra, Shahin Jabbari, Steven Wu, and Himabindu Lakkaraju. 2022. “The Disagreement Problem in Explainable Machine Learning: A Practitioner’s Perspective.” arXiv. https://doi.org/10.48550/ARXIV.2202.01602.
Krivoruchko, Konstantin, and Alexander Gribov. 2019. “Evaluation of Empirical Bayesian Kriging.” Spatial Statistics 32 (August). https://doi.org/10.1016/j.spasta.2019.100368.
Krizhevsky, Alex. 2014. “One Weird Trick for Parallelizing Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1404.5997.
Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. 2017. “ImageNet Classification with Deep Convolutional Neural Networks.” Communications of the ACM 60 (May). https://doi.org/10.1145/3065386.
Kroon, Dirk-Jan. n.d. “Segmentation of the Mandibular Canal in Cone-Beam CT Data.” https://doi.org/10.3990/1.9789036532808.
Kruglov, Alexey. 2018. “Channel-Wise Pruning of Neural Networks with Tapering Resource Constraint.” arXiv. https://doi.org/10.48550/ARXIV.1812.07060.
Krylov, N. V., and M. Röckner. 2004. “Strong Solutions of Stochastic Equations with Singular Time Dependent Drift.” Probability Theory and Related Fields 131 (May). https://doi.org/10.1007/s00440-004-0361-z.
Kuba, Jakub Grudzien, Muning Wen, Yaodong Yang, Linghui Meng, Shangding Gu, Haifeng Zhang, David Henry Mguni, and Jun Wang. 2021. “Settling the Variance of Multi-Agent Policy Gradients.” arXiv. https://doi.org/10.48550/ARXIV.2108.08612.
Kuhn, Lorenz, Yarin Gal, and Sebastian Farquhar. 2023. “Semantic Uncertainty: Linguistic Invariances for Uncertainty Estimation in Natural Language Generation.” arXiv. https://doi.org/10.48550/ARXIV.2302.09664.
Kulis, Brian, and Michael I. Jordan. 2011. “Revisiting k-Means: New Algorithms via Bayesian Nonparametrics.” arXiv. https://doi.org/10.48550/ARXIV.1111.0352.
Kumar, Ankit, Ozan Irsoy, Peter Ondruska, Mohit Iyyer, James Bradbury, Ishaan Gulrajani, Victor Zhong, Romain Paulus, and Richard Socher. 2015. “Ask Me Anything: Dynamic Memory Networks for Natural Language Processing.” arXiv. https://doi.org/10.48550/ARXIV.1506.07285.
Kumar, Ravi, Andrew Tomkins, Sergei Vassilvitskii, and Erik Vee. 2015. “Inverting a Steady-State.” Proceedings of the Eighth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/2684822.2685310.
Kumar, Srijan, Xikun Zhang, and Jure Leskovec. 2019. “Predicting Dynamic Embedding Trajectory in Temporal Interaction Networks.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330895.
Kumar, Varun, Ashutosh Choudhary, and Eunah Cho. 2020. “Data Augmentation Using Pre-Trained Transformer Models,” March. http://arxiv.org/abs/2003.02245v2.
Kumarage, Tharindu, Joshua Garland, Amrita Bhattacharjee, Kirill Trapeznikov, Scott Ruston, and Huan Liu. 2023. “Stylometric Detection of AI-Generated Text in Twitter Timelines.” arXiv. https://doi.org/10.48550/ARXIV.2303.03697.
Kumawat, Sudhakar, and Shanmuganathan Raman. 2019. “LP-3DCNN: Unveiling Local Phase in 3D Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1904.03498.
Kung. 1982. “Why Systolic Architectures?” Computer 15 (January). https://doi.org/10.1109/mc.1982.1653825.
Kurach, Karol, Marcin Andrychowicz, and Ilya Sutskever. 2015. “Neural Random-Access Machines.” arXiv. https://doi.org/10.48550/ARXIV.1511.06392.
Kurakin, Alexey, Ian Goodfellow, Samy Bengio, Yinpeng Dong, Fangzhou Liao, Ming Liang, Tianyu Pang, et al. 2018. “Adversarial Attacks and Defences Competition.” arXiv. https://doi.org/10.48550/ARXIV.1804.00097.
Kurian, N., J. M. Cherian, N. A. Sudharson, K. G. Varghese, and S. Wadhwa. 2023. “AI Is Now Everywhere.” British Dental Journal 234 (January). https://doi.org/10.1038/s41415-023-5461-1.
Kusner, Matt J., and José Miguel Hernández-Lobato. 2016. “GANS for Sequences of Discrete Elements with the Gumbel-Softmax Distribution.” arXiv. https://doi.org/10.48550/ARXIV.1611.04051.
Kuzborskij, Ilja, and Francesco Orabona. 2016. “Fast Rates by Transferring from Auxiliary Hypotheses.” Machine Learning 106 (October). https://doi.org/10.1007/s10994-016-5594-4.
Kuzman, Taja, Igor Mozetič, and Nikola Ljubešić. 2023. “ChatGPT: Beginning of an End of Manual Linguistic Data Annotation? Use Case of Automatic Genre Identification.” arXiv. https://doi.org/10.48550/ARXIV.2303.03953.
Kwon, Minae, Sang Michael Xie, Kalesha Bullard, and Dorsa Sadigh. 2023. “Reward Design with Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.00001.
Kwon, Woosuk, Zhuohan Li, Siyuan Zhuang, Ying Sheng, Lianmin Zheng, Cody Hao Yu, Joseph Gonzalez, Hao Zhang, and Ion Stoica. 2023. “Efficient Memory Management for Large Language Model Serving with PagedAttention.” Proceedings of the 29th Symposium on Operating Systems Principles, October. https://doi.org/10.1145/3600006.3613165.
Labrak, Yanis, Adrien Bazoge, Richard Dufour, Mickael Rouvier, Emmanuel Morin, Béatrice Daille, and Pierre-Antoine Gourraud. 2023a. “FrenchMedMCQA: A French Multiple-Choice Question Answering Dataset for Medical Domain.” arXiv. https://doi.org/10.48550/ARXIV.2304.04280.
———. 2023b. “DrBERT: A Robust Pre-Trained Model in French for Biomedical and Clinical Domains,” April. https://doi.org/10.1101/2023.04.03.535368.
Lacoste-Julien, Simon, Martin Jaggi, Mark Schmidt, and Patrick Pletscher. 2012. “Block-Coordinate Frank-Wolfe Optimization for Structural SVMs.” arXiv. https://doi.org/10.48550/ARXIV.1207.4747.
Lacroix, Timothée, Nicolas Usunier, and Guillaume Obozinski. 2018. “Canonical Tensor Decomposition for Knowledge Base Completion.” arXiv. https://doi.org/10.48550/ARXIV.1806.07297.
Lai, Guokun, Wei-Cheng Chang, Yiming Yang, and Hanxiao Liu. 2017. “Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1703.07015.
Laine, Samuli, and Timo Aila. 2016. “Temporal Ensembling for Semi-Supervised Learning.” arXiv. https://doi.org/10.48550/ARXIV.1610.02242.
Lakhmiri, Dounia, Sébastien Le Digabel, and Christophe Tribes. 2019. “HyperNOMAD: Hyperparameter Optimization of Deep Neural Networks Using Mesh Adaptive Direct Search.” arXiv. https://doi.org/10.48550/ARXIV.1907.01698.
Lam, Hoang Thanh, Johann-Michael Thiebaut, Mathieu Sinn, Bei Chen, Tiep Mai, and Oznur Alkan. 2017. “One Button Machine for Automating Feature Engineering in Relational Databases.” arXiv. https://doi.org/10.48550/ARXIV.1706.00327.
Lamb, Alex, Anirudh Goyal, Ying Zhang, Saizheng Zhang, Aaron Courville, and Yoshua Bengio. 2016. “Professor Forcing: A New Algorithm for Training Recurrent Networks.” arXiv. https://doi.org/10.48550/ARXIV.1610.09038.
Lample, Guillaume, and Devendra Singh Chaplot. 2016. “Playing FPS Games with Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1609.05521.
Lample, Guillaume, and François Charton. 2019. “Deep Learning for Symbolic Mathematics.” arXiv. https://doi.org/10.48550/ARXIV.1912.01412.
Lample, Guillaume, Myle Ott, Alexis Conneau, Ludovic Denoyer, and Marc’Aurelio Ranzato. 2018. “Phrase-Based &Amp; Neural Unsupervised Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1804.07755.
Lan, Guanghui, Soomin Lee, and Yi Zhou. 2017. “Communication-Efficient Algorithms for Decentralized and Stochastic Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1701.03961.
Lan, Li-Cheng, Wei Li, Ting-Han Wei, and I-Chen Wu. 2019. “Multiple Policy Value Monte Carlo Tree Search.” arXiv. https://doi.org/10.48550/ARXIV.1905.13521.
Lan, Zhenzhong, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. “ALBERT: A Lite BERT for Self-Supervised Learning of Language Representations.” arXiv. https://doi.org/10.48550/ARXIV.1909.11942.
Lanctot, Marc, Vinicius Zambaldi, Audrunas Gruslys, Angeliki Lazaridou, Karl Tuyls, Julien Perolat, David Silver, and Thore Graepel. 2017. “A Unified Game-Theoretic Approach to Multiagent Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1711.00832.
Landsberg, J. M., and Zach Teitler. 2009. “On the Ranks and Border Ranks of Symmetric Tensors.” arXiv. https://doi.org/10.48550/ARXIV.0901.0487.
Lang, Stefan, and Andreas Brezger. 2004. “Bayesian p-Splines.” Journal of Computational and Graphical Statistics 13 (March). https://doi.org/10.1198/1061860043010.
Langford, John, Lihong Li, and Tong Zhang. 2008. “Sparse Online Learning via Truncated Gradient.” arXiv. https://doi.org/10.48550/ARXIV.0806.4686.
Langley, Pat, and Stephanie Sage. 2013. “Induction of Selective Bayesian Classifiers.” arXiv. https://doi.org/10.48550/ARXIV.1302.6828.
Langroudi, Hamed F., Vedant Karia, Tej Pandit, and Dhireesha Kudithipudi. 2021. “TENT: Efficient Quantization of Neural Networks on the Tiny Edge with Tapered FixEd PoiNT.” arXiv. https://doi.org/10.48550/ARXIV.2104.02233.
Laptev, Nikolay, Saeed Amizadeh, and Ian Flint. 2015. “Generic and Scalable Framework for Automated Time-Series Anomaly Detection.” Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2783258.2788611.
Laradji, Issam, and Reza Babanezhad. 2018. “M-ADDA: Unsupervised Domain Adaptation with Deep Metric Learning.” arXiv. https://doi.org/10.48550/ARXIV.1807.02552.
Lathuiliere, Stephane, Pablo Mesejo, Xavier Alameda-Pineda, and Radu Horaud. 2020. “A Comprehensive Analysis of Deep Regression.” IEEE Transactions on Pattern Analysis and Machine Intelligence 42 (September). https://doi.org/10.1109/tpami.2019.2910523.
Latif, Ehsan, Gengchen Mai, Matthew Nyaaba, Xuansheng Wu, Ninghao Liu, Guoyu Lu, Sheng Li, Tianming Liu, and Xiaoming Zhai. 2023. “AGI: Artificial General Intelligence for Education.” arXiv. https://doi.org/10.48550/ARXIV.2304.12479.
Latif, Siddique, Aun Zaidi, Heriberto Cuayahuitl, Fahad Shamshad, Moazzam Shoukat, and Junaid Qadir. 2023. “Transformers in Speech Processing: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2303.11607.
Lau, Ada, and Patrick McSharry. 2010. “Approaches for Multi-Step Density Forecasts with Application to Aggregated Wind Power.” The Annals of Applied Statistics 4 (September). https://doi.org/10.1214/09-aoas320.
Laube, Kevin Alexander, and Andreas Zell. 2019. “Prune and Replace NAS.” 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), December. https://doi.org/10.1109/icmla.2019.00158.
Lavin, Alexander, and Subutai Ahmad. 2015. “Evaluating Real-Time Anomaly Detection Algorithms – the Numenta Anomaly Benchmark.” 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), December. https://doi.org/10.1109/icmla.2015.141.
Lavine, Isaac, Michael Lindon, and Mike West. 2021. “Adaptive Variable Selection for Sequential Prediction in Multivariate Dynamic Models.” Bayesian Analysis 16 (December). https://doi.org/10.1214/20-ba1245.
Lavine, Michael, and Mike West. 1992. “A Bayesian Method for Classification and Discrimination.” Canadian Journal of Statistics 20 (December). https://doi.org/10.2307/3315614.
Lawler, Gregory F., and Vlada Limic. 2010. “Random Walk: A Modern Introduction,” June. https://doi.org/10.1017/cbo9780511750854.
Lawler, Gregory F., and Wendelin Werner. 2004. “The Brownian Loop Soup.” Probability Theory and Related Fields 128 (January). https://doi.org/10.1007/s00440-003-0319-6.
Lawrence, Michael, and Martin Morgan. 2014. “Scalable Genomics with r and Bioconductor.” Statistical Science 29 (May). https://doi.org/10.1214/14-sts476.
Le, Hoang M., Yisong Yue, Peter Carr, and Patrick Lucey. 2017. “Coordinated Multi-Agent Imitation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1703.03121.
Le, Matt, Stephen Roller, Laetitia Papaxanthos, Douwe Kiela, and Maximilian Nickel. 2019. “Inferring Concept Hierarchies from Text Corpora via Hyperbolic Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1902.00913.
Le, Nhat, Thang Pham, Tuong Do, Erman Tjiputra, Quang D. Tran, and Anh Nguyen. 2023. “Music-Driven Group Choreography.” arXiv. https://doi.org/10.48550/ARXIV.2303.12337.
Le, Quoc Viet, Tamas Sarlos, and Alexander Johannes Smola. 2014. “Fastfood: Approximate Kernel Expansions in Loglinear Time.” arXiv. https://doi.org/10.48550/ARXIV.1408.3060.
Le, Quoc V., and Tomas Mikolov. 2014. “Distributed Representations of Sentences and Documents.” arXiv. https://doi.org/10.48550/ARXIV.1405.4053.
Le, Quoc V., Marc’Aurelio Ranzato, Rajat Monga, Matthieu Devin, Kai Chen, Greg S. Corrado, Jeff Dean, and Andrew Y. Ng. 2011. “Building High-Level Features Using Large Scale Unsupervised Learning.” arXiv. https://doi.org/10.48550/ARXIV.1112.6209.
Le, Truc, and Ye Duan. 2018. “PointGrid: A Deep Network for 3D Shape Understanding.” 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, June. https://doi.org/10.1109/cvpr.2018.00959.
Le, Van-Hoang, and Hongyu Zhang. 2023. “Log Parsing with Prompt-Based Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.2302.07435.
LeClair, Alexander, Sakib Haque, Lingfei Wu, and Collin McMillan. 2020. “Improved Code Summarization via a Graph Neural Network.” Proceedings of the 28th International Conference on Program Comprehension, July. https://doi.org/10.1145/3387904.3389268.
LeCun, Yann, Koray Kavukcuoglu, and Clement Farabet. 2010. “Convolutional Networks and Applications in Vision.” Proceedings of 2010 IEEE International Symposium on Circuits and Systems, May. https://doi.org/10.1109/iscas.2010.5537907.
Lecun, Y., L. Bottou, Y. Bengio, and P. Haffner. 1998. “Gradient-Based Learning Applied to Document Recognition.” Proceedings of the IEEE 86. https://doi.org/10.1109/5.726791.
Ledoit, Olivier, and Sandrine Péché. 2010. “Eigenvectors of Some Large Sample Covariance Matrix Ensembles.” Probability Theory and Related Fields 151 (May). https://doi.org/10.1007/s00440-010-0298-3.
Lee, Ariel N., Cole J. Hunter, and Nataniel Ruiz. 2023. “Platypus: Quick, Cheap, and Powerful Refinement of LLMs.” arXiv. https://doi.org/10.48550/ARXIV.2308.07317.
Lee, Chung-Wei, Wei Fang, Chih-Kuan Yeh, and Yu-Chiang Frank Wang. 2017. “Multi-Label Zero-Shot Learning with Structured Knowledge Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1711.06526.
Lee, Dongsoo, Se Jung Kwon, Byeongwook Kim, Parichay Kapoor, and Gu-Yeon Wei. 2019. “Network Pruning for Low-Rank Binary Indexing.” arXiv. https://doi.org/10.48550/ARXIV.1905.05686.
Lee, Jason, Kyunghyun Cho, and Thomas Hofmann. 2016. “Fully Character-Level Neural Machine Translation Without Explicit Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1610.03017.
Lee, Jongseok, Matthias Humt, Jianxiang Feng, and Rudolph Triebel. 2020. “Estimating Model Uncertainty of Neural Networks in Sparse Information Form.” arXiv. https://doi.org/10.48550/ARXIV.2006.11631.
Lee, Junhyun, Inyeop Lee, and Jaewoo Kang. 2019. “Self-Attention Graph Pooling.” arXiv. https://doi.org/10.48550/ARXIV.1904.08082.
Lee, Kang-Ho, JoonHyun Jeong, and Sung-Ho Bae. 2019. “An Inter-Layer Weight Prediction and Quantization for Deep Neural Networks Based on a Smoothly Varying Weight Hypothesis.” arXiv. https://doi.org/10.48550/ARXIV.1907.06835.
Lee, Kenton, Mike Lewis, and Luke Zettlemoyer. 2016. “Global Neural CCG Parsing with Optimality Guarantees.” arXiv. https://doi.org/10.48550/ARXIV.1607.01432.
Lee, Kenton, Shimi Salant, Tom Kwiatkowski, Ankur Parikh, Dipanjan Das, and Jonathan Berant. 2016. “Learning Recurrent Span Representations for Extractive Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1611.01436.
Lee, Kimin, Kibok Lee, Honglak Lee, and Jinwoo Shin. 2018. “A Simple Unified Framework for Detecting Out-of-Distribution Samples and Adversarial Attacks.” arXiv. https://doi.org/10.48550/ARXIV.1807.03888.
Lee, Kimin, Hao Liu, Moonkyung Ryu, Olivia Watkins, Yuqing Du, Craig Boutilier, Pieter Abbeel, Mohammad Ghavamzadeh, and Shixiang Shane Gu. 2023. “Aligning Text-to-Image Models Using Human Feedback.” arXiv. https://doi.org/10.48550/ARXIV.2302.12192.
Lee, Meng-Chieh, Yue Zhao, Aluna Wang, Pierre Jinghong Liang, Leman Akoglu, Vincent S. Tseng, and Christos Faloutsos. 2020. “AutoAudit: Mining Accounting and Time-Evolving Graphs.” arXiv. https://doi.org/10.48550/ARXIV.2011.00447.
Lee, Seyoung, Sunmin Lee, Yongwoo Lee, and Jehee Lee. 2021. “Learning a Family of Motor Skills from a Single Motion Clip.” ACM Transactions on Graphics 40 (July). https://doi.org/10.1145/3450626.3459774.
Lee, Sungjin. 2017. “Toward Continual Learning for Conversational Agents.” arXiv. https://doi.org/10.48550/ARXIV.1712.09943.
Lee, Tai Sing, and David Mumford. 2003. “Hierarchical Bayesian Inference in the Visual Cortex.” Journal of the Optical Society of America A 20 (July). https://doi.org/10.1364/josaa.20.001434.
Lee, Wenke, Salvatore Stolfo, and Philip K. Chan. 1997. “Learning Patterns from Unix Process Execution Traces for Intrusion Detection.” Columbia University. https://doi.org/10.7916/D8B56RF2.
Leek, Jeffrey T., and Roger D. Peng. 2015. “Statistics: P Values Are Just the Tip of the Iceberg.” Nature 520 (April). https://doi.org/10.1038/520612a.
Legg, Shane, and Marcus Hutter. 2007. “Universal Intelligence: A Definition of Machine Intelligence.” arXiv. https://doi.org/10.48550/ARXIV.0712.3329.
Lei, Jiahui, Congyue Deng, Karl Schmeckpeper, Leonidas Guibas, and Kostas Daniilidis. 2023. “EFEM: Equivariant Neural Field Expectation Maximization for 3D Object Segmentation Without Scene Supervision.” arXiv. https://doi.org/10.48550/ARXIV.2303.15440.
Leibo, Joel Z., Vinicius Zambaldi, Marc Lanctot, Janusz Marecki, and Thore Graepel. 2017. “Multi-Agent Reinforcement Learning in Sequential Social Dilemmas.” arXiv. https://doi.org/10.48550/ARXIV.1702.03037.
Lenselink, Eelke B., Niels ten Dijke, Brandon Bongers, George Papadatos, Herman W. T. van Vlijmen, Wojtek Kowalczyk, Adriaan P. IJzerman, and Gerard J. P. van Westen. 2017. “Beyond the Hype: Deep Neural Networks Outperform Established Methods Using a ChEMBL Bioactivity Benchmark Set.” Journal of Cheminformatics 9 (August). https://doi.org/10.1186/s13321-017-0232-0.
Leroux, Sam, Pieter Simoens, Meelis Lootus, Kartik Thakore, and Akshay Sharma. 2022. “TinyMLOps: Operational Challenges for Widespread Edge AI Adoption.” arXiv. https://doi.org/10.48550/ARXIV.2203.10923.
Lesage, Jonathan, Robert Brennan, Sarah Elaine Eaton, Beatriz Moya, Brenda McDermott, Jason Wiens, and Kai Herrero. 2023. “Exploring Natural Language Processing in Mechanical Engineering Education: Implications for Academic Integrity.” International Journal of Mechanical Engineering Education 52 (March). https://doi.org/10.1177/03064190231166665.
Leskovec, Jure, Deepayan Chakrabarti, Jon Kleinberg, Christos Faloutsos, and Zoubin Ghahramani. 2008. “Kronecker Graphs: An Approach to Modeling Networks.” arXiv. https://doi.org/10.48550/ARXIV.0812.4905.
Lester, Brian, Rami Al-Rfou, and Noah Constant. 2021. “The Power of Scale for Parameter-Efficient Prompt Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2104.08691.
Levi, Evgeny, and Radu V. Craiu. 2022. “Finding Our Way in the Dark: Approximate MCMC for Approximate Bayesian Methods.” Bayesian Analysis 17 (March). https://doi.org/10.1214/20-ba1250.
Levine, David M, Rudraksh Tuwani, Benjamin Kompa, Amita Varma, Samuel G. Finlayson, Ateev Mehrotra, and Andrew Beam. 2023. “The Diagnostic and Triage Accuracy of the GPT-3 Artificial Intelligence Model,” February. https://doi.org/10.1101/2023.01.30.23285067.
Levinkov, Evgeny, Jonas Uhrig, Siyu Tang, Mohamed Omran, Eldar Insafutdinov, Alexander Kirillov, Carsten Rother, Thomas Brox, Bernt Schiele, and Bjoern Andres. 2016. “Joint Graph Decomposition and Node Labeling: Problem, Algorithms, Applications.” arXiv. https://doi.org/10.48550/ARXIV.1611.04399.
Lévy-Leduc, Céline, and François Roueff. 2009. “Detection and Localization of Change-Points in High-Dimensional Network Traffic Data.” The Annals of Applied Statistics 3 (June). https://doi.org/10.1214/08-aoas232.
Lewicki, Michael S. 1994. “Bayesian Modeling and Classification of Neural Signals.” Neural Computation 6 (September). https://doi.org/10.1162/neco.1994.6.5.1005.
Lewin, Alex, Natalia Bochkina, and Sylvia Richardson. 2007. “Fully Bayesian Mixture Model for Differential Gene Expression: Simulations and Model Checks.” Statistical Applications in Genetics and Molecular Biology 6 (January). https://doi.org/10.2202/1544-6115.1314.
Lewis, David D., and William A. Gale. 1994. “A Sequential Algorithm for Training Text Classifiers.” arXiv. https://doi.org/10.48550/ARXIV.CMP-LG/9407020.
Li, Aaron Q., Amr Ahmed, Sujith Ravi, and Alexander J. Smola. 2014. “Reducing the Sampling Complexity of Topic Models.” Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2623330.2623756.
Li, Ao, Jemin Andrew Choi, and Fan Long. 2020. “Securing Smart Contract with Runtime Validation.” Proceedings of the 41st ACM SIGPLAN Conference on Programming Language Design and Implementation, June. https://doi.org/10.1145/3385412.3385982.
Li, Bin, and Steven C. H. Hoi. 2012. “Online Portfolio Selection: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.1212.2129.
Li, Bo, Gexiang Fang, Yang Yang, Quansen Wang, Wei Ye, Wen Zhao, and Shikun Zhang. 2023. “Evaluating ChatGPT’s Information Extraction Capabilities: An Assessment of Performance, Explainability, Calibration, and Faithfulness.” arXiv. https://doi.org/10.48550/ARXIV.2304.11633.
Li, Bohan, Hao Zhou, Junxian He, Mingxuan Wang, Yiming Yang, and Lei Li. 2020. “On the Sentence Embeddings from Pre-Trained Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2011.05864.
Li, Bo, Yuanhan Zhang, Liangyu Chen, Jinghao Wang, Jingkang Yang, and Ziwei Liu. 2023. “Otter: A Multi-Modal Model with in-Context Instruction Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2305.03726.
Li, Chao, Zhiyuan Liu, Mengmeng Wu, Yuchi Xu, Pipei Huang, Huan Zhao, Guoliang Kang, Qiwei Chen, Wei Li, and Dik Lun Lee. 2019. “Multi-Interest Network with Dynamic Routing for Recommendation at Tmall.” arXiv. https://doi.org/10.48550/ARXIV.1904.08030.
Li, Chao, Xiaokong Ma, Bing Jiang, Xiangang Li, Xuewei Zhang, Xiao Liu, Ying Cao, Ajay Kannan, and Zhenyao Zhu. 2017. “Deep Speaker: An End-to-End Neural Speaker Embedding System.” arXiv. https://doi.org/10.48550/ARXIV.1705.02304.
Li, Chuan, and Michael Wand. 2016. “Precomputed Real-Time Texture Synthesis with Markovian Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1604.04382.
Li, Chunhui, Xingshu Chen, Haizhou Wang, Yu Zhang, and Peiming Wang. 2020. “An End-to-End Attack on Text-Based CAPTCHAs Based on Cycle-Consistent Generative Adversarial Network.” arXiv. https://doi.org/10.48550/ARXIV.2008.11603.
Li, Conglong, Minjia Zhang, David G. Andersen, and Yuxiong He. 2020. “Improving Approximate Nearest Neighbor Search Through Learned Adaptive Early Termination.” Proceedings of the 2020 ACM SIGMOD International Conference on Management of Data, May. https://doi.org/10.1145/3318464.3380600.
Li, Daoyuan, Jessica Lin, Tegawendé Bissyandé, Jacques Klein, and Yves Le Traon. 2018. “Extracting Statistical Graph Features for Accurate and Efficient Time Series Classification.” https://doi.org/10.5441/002/EDBT.2018.19.
Li, En, Zhi Zhou, and Xu Chen. 2018. “Edge Intelligence: On-Demand Deep Learning Model Co-Inference with Device-Edge Synergy.” arXiv. https://doi.org/10.48550/ARXIV.1806.07840.
Li, Feng, Zhenrui Chen, Pengjie Wang, Yi Ren, Di Zhang, and Xiaoyu Zhu. 2019. “Graph Intention Network for Click-Through Rate Prediction in Sponsored Search.” Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3331184.3331283.
———. 2021. “Graph Intention Network for Click-Through Rate Prediction in Sponsored Search.” arXiv. https://doi.org/10.48550/ARXIV.2103.16164.
Li, Fengfu, Bin Liu, Xiaoxing Wang, Bo Zhang, and Junchi Yan. 2016. “Ternary Weight Networks.” arXiv. https://doi.org/10.48550/ARXIV.1605.04711.
LI, Hang. 2011. “A Short Introduction to Learning to Rank.” IEICE Transactions on Information and Systems E94-D. https://doi.org/10.1587/transinf.e94.d.1854.
Li, Hang, Ahmed Mourad, Shengyao Zhuang, Bevan Koopman, and Guido Zuccon. 2021. “Pseudo Relevance Feedback with Deep Language Models and Dense Retrievers: Successes and Pitfalls.” arXiv. https://doi.org/10.48550/ARXIV.2108.11044.
Li, Hao, Asim Kadav, Igor Durdanovic, Hanan Samet, and Hans Peter Graf. 2016. “Pruning Filters for Efficient ConvNets.” arXiv. https://doi.org/10.48550/ARXIV.1608.08710.
Li, Hao, Han Liu, Dewei Hu, Jiacheng Wang, and Ipek Oguz. 2023. “Promise:prompt-Driven 3D Medical Image Segmentation Using Pretrained Image Foundation Models.” arXiv. https://doi.org/10.48550/ARXIV.2310.19721.
Li, Haoyang, Jing Zhang, Cuiping Li, and Hong Chen. 2023. “RESDSQL: Decoupling Schema Linking and Skeleton Parsing for Text-to-SQL.” arXiv. https://doi.org/10.48550/ARXIV.2302.05965.
Li, Jia, Zhichao Han, Hong Cheng, Jiao Su, Pengyun Wang, Jianfeng Zhang, and Lujia Pan. 2019. “Predicting Path Failure in Time-Evolving Graphs.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330847.
Li, Jiangmeng, Wenyi Mo, Wenwen Qiang, Bing Su, and Changwen Zheng. 2022. “Supporting Vision-Language Model Inference with Causality-Pruning Knowledge Prompt.” arXiv. https://doi.org/10.48550/ARXIV.2205.11100.
Li, Jiwei, and Dan Jurafsky. 2015. “Do Multi-Sense Embeddings Improve Natural Language Understanding?” arXiv. https://doi.org/10.48550/ARXIV.1506.01070.
Li, Jundong, Harsh Dani, Xia Hu, Jiliang Tang, Yi Chang, and Huan Liu. 2017. “Attributed Network Embedding for Learning in a Dynamic Environment.” Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3132847.3132919.
Li, Junjie, Yixin Zhang, Zilei Wang, Keyu Tu, and Saihui Hou. 2021. “Probabilistic Contrastive Learning for Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.2111.06021.
Li, Junnan, Dongxu Li, Silvio Savarese, and Steven Hoi. 2023. “BLIP-2: Bootstrapping Language-Image Pre-Training with Frozen Image Encoders and Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.12597.
Li, Junyi, Jie Chen, Ruiyang Ren, Xiaoxue Cheng, Wayne Xin Zhao, Jian-Yun Nie, and Ji-Rong Wen. 2024. “The Dawn After the Dark: An Empirical Study on Factuality Hallucination in Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.03205.
Li, Junyi, Xiaoxue Cheng, Wayne Xin Zhao, Jian-Yun Nie, and Ji-Rong Wen. 2023. “HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2305.11747.
Li, Ke, and Jitendra Malik. 2016. “Learning to Optimize.” arXiv. https://doi.org/10.48550/ARXIV.1606.01885.
Li, Kunchang, Yali Wang, Yizhuo Li, Yi Wang, Yinan He, Limin Wang, and Yu Qiao. 2023. “Unmasked Teacher: Towards Training-Efficient Video Foundation Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.16058.
Li, Lantian, Dong Wang, Zhiyong Zhang, and Thomas Fang Zheng. 2015. “Deep Speaker Vectors for Semi Text-Independent Speaker Verification.” arXiv. https://doi.org/10.48550/ARXIV.1505.06427.
Li, Lei, Yongfeng Zhang, and Li Chen. 2022. “Personalized Prompt Learning for Explainable Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.2202.07371.
———. 2023. “Prompt Distillation for Efficient LLM-Based Recommendation.” Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3583780.3615017.
Li, Ling. 2006. “Multiclass Boosting with Repartitioning.” Proceedings of the 23rd International Conference on Machine Learning - ICML ’06. https://doi.org/10.1145/1143844.1143916.
Li, Lisha, Kevin Jamieson, Giulia DeSalvo, Afshin Rostamizadeh, and Ameet Talwalkar. 2016. “Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1603.06560.
Li, Minghao, Xi He, Wendong Zhang, Shuyang Qu, Lulu Rodriguez, and James M. Gbeda. 2023. “Farmers’ Reactions to the US–China Trade War: Perceptions Versus Behaviors.” Journal of the Agricultural and Applied Economics Association 2 (June). https://doi.org/10.1002/jaa2.68.
Li, Mingxuan, Yazhe Wang, Shuai Ma, Chao Liu, Dongdong Huo, Yu Wang, and Zhen Xu. 2023. “Auto-Tuning with Reinforcement Learning for Permissioned Blockchain Systems.” Proceedings of the VLDB Endowment 16 (January). https://doi.org/10.14778/3579075.3579076.
Li, Peipei, Haixiang Zhang, Xuegang Hu, and Xindong Wu. 2022. “High-Dimensional Multi-Label Data Stream Classification with Concept Drifting Detection.” IEEE Transactions on Knowledge and Data Engineering. https://doi.org/10.1109/tkde.2022.3200068.
Li, Peng, Jiaxin Mi, and Jiaxin Mi. 2022. “Event Detection with Dual Relational Graph Attention Networks.” https://doi.org/10.48448/62FC-GT66.
Li, Ping. 2009. “ABC-LogitBoost for Multi-Class Classification.” arXiv. https://doi.org/10.48550/ARXIV.0908.4144.
———. 2012. “Robust LogitBoost and Adaptive Base Class (ABC) LogitBoost.” arXiv. https://doi.org/10.48550/ARXIV.1203.3491.
Li, Qian, Jianxin Li, Jiawei Sheng, Shiyao Cui, Jia Wu, Yiming Hei, Hao Peng, et al. 2021. “A Survey on Deep Learning Event Extraction: Approaches and Applications.” arXiv. https://doi.org/10.48550/ARXIV.2107.02126.
Li, Qinbin, Yiqun Diao, Quan Chen, and Bingsheng He. 2021. “Federated Learning on Non-IID Data Silos: An Experimental Study.” arXiv. https://doi.org/10.48550/ARXIV.2102.02079.
Li, Qiujia, Yu Zhang, Bo Li, Liangliang Cao, and Philip C. Woodland. 2021. “Residual Energy-Based Models for End-to-End Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2103.14152.
Li, Raymond, Loubna Ben Allal, Yangtian Zi, Niklas Muennighoff, Denis Kocetkov, Chenghao Mou, Marc Marone, et al. 2023. “StarCoder: May the Source Be with You!” arXiv. https://doi.org/10.48550/ARXIV.2305.06161.
Li, Ruihui, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, and Pheng-Ann Heng. 2019. “PU-GAN: A Point Cloud Upsampling Adversarial Network.” arXiv. https://doi.org/10.48550/ARXIV.1907.10844.
Li, Ruiyu, Makarand Tapaswi, Renjie Liao, Jiaya Jia, Raquel Urtasun, and Sanja Fidler. 2017. “Situation Recognition with Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1708.04320.
Li, Sai, Tianxi Cai, and Rui Duan. 2021. “Targeting Underrepresented Populations in Precision Medicine: A Federated Transfer Learning Approach.” arXiv. https://doi.org/10.48550/ARXIV.2108.12112.
Li, Sheng, Jongsoo Park, and Ping Tak Peter Tang. 2017. “Enabling Sparse Winograd Convolution by Native Pruning.” arXiv. https://doi.org/10.48550/ARXIV.1702.08597.
Li, Shi-Jie, Ming-Ming Cheng, Yun Liu, Shao-Ping Lu, YaHui Wang, and Victor Adrian Prisacariu. 2018. “Structured Skip List: A Compact Data Structure for 3D Reconstruction.” 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), October. https://doi.org/10.1109/iros.2018.8594075.
Li, Shuai, Zhao Song, Yu Xia, Tong Yu, and Tianyi Zhou. 2023. “The Closeness of in-Context Learning and Weight Shifting for Softmax Regression.” arXiv. https://doi.org/10.48550/ARXIV.2304.13276.
Li, Sihang, Zhiyuan Liu, Yanchen Luo, Xiang Wang, Xiangnan He, Kenji Kawaguchi, Tat-Seng Chua, and Qi Tian. 2024. “Towards 3D Molecule-Text Interpretation in Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.13923.
Li, Wen, Limin Wang, Wei Li, Eirikur Agustsson, and Luc Van Gool. 2017. “WebVision Database: Visual Learning and Understanding from Web Data.” arXiv. https://doi.org/10.48550/ARXIV.1708.02862.
Li, Xiang, Shuo Chen, Xiaolin Hu, and Jian Yang. 2018. “Understanding the Disharmony Between Dropout and Batch Normalization by Variance Shift.” arXiv. https://doi.org/10.48550/ARXIV.1801.05134.
Li, Xiang, Changhe Song, Jingbei Li, Zhiyong Wu, Jia Jia, and Helen Meng. 2021. “Towards Multi-Scale Style Control for Expressive Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.2104.03521.
Li, Xiang, Luke Vilnis, and Andrew McCallum. 2017. “Improved Representation Learning for Predicting Commonsense Ontologies,” August. http://arxiv.org/abs/1708.00549v1.
Li, Xiang, Wenhai Wang, Xiaolin Hu, and Jian Yang. 2019. “Selective Kernel Networks.” arXiv. https://doi.org/10.48550/ARXIV.1903.06586.
Li, Xinglin, Jiajing Chen, Jinhui Ouyang, Hanhui Deng, Senem Velipasalar, and Di Wu. 2023. “ToThePoint: Efficient Contrastive Learning of 3D Point Clouds via Recycling.” 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52729.2023.02086.
Li, Xin, Zequn Jie, Jiashi Feng, Changsong Liu, and Shuicheng Yan. 2017. “Learning with Rethinking: Recurrently Improving Convolutional Neural Networks Through Feedback.” arXiv. https://doi.org/10.48550/ARXIV.1708.04483.
Li, Xinke, Henghui Ding, Zekun Tong, Yuwei Wu, and Yeow Meng Chee. 2022. “Primitive3D: 3D Object Dataset Synthesis from Randomly Assembled Primitives.” arXiv. https://doi.org/10.48550/ARXIV.2205.12627.
Li, Xin-Yi, Wei-Jun Lei, and Yu-Bin Yang. 2022. “From Easy to Hard: Two-Stage Selector and Reader for Multi-Hop Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.2205.11729.
Li, Xiujun, Yun-Nung Chen, Lihong Li, Jianfeng Gao, and Asli Celikyilmaz. 2017. “End-to-End Task-Completion Neural Dialogue Systems.” arXiv. https://doi.org/10.48550/ARXIV.1703.01008.
Li, Yanghao, Naiyan Wang, Jianping Shi, Jiaying Liu, and Xiaodi Hou. 2016. “Revisiting Batch Normalization for Practical Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1603.04779.
Li, Yang, Yu Shen, Huaijun Jiang, Wentao Zhang, Zhi Yang, Ce Zhang, and Bin Cui. 2022. “TransBO: Hyperparameter Optimization via Two-Phase Transfer Learning.” Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3534678.3539255.
Li, Yang, and Thomas Strohmer. 2019. “What Happens on the Edge, Stays on the Edge: Toward Compressive Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1909.01539.
Li, Yanzeng, Jiangxia Cao, Xin Cong, Zhenyu Zhang, Bowen Yu, Hongsong Zhu, and Tingwen Liu. 2022. “Enhancing Chinese Pre-Trained Language Model via Heterogeneous Linguistics Graph.” Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/2022.acl-long.140.
Li, Yifu, Ran Jin, and Yuan Luo. 2018. “Classifying Relations in Clinical Narratives Using Segment Graph Convolutional and Recurrent Neural Networks (Seg-GCRNs).” Journal of the American Medical Informatics Association 26 (December). https://doi.org/10.1093/jamia/ocy157.
Li, Yikuan, Ramsey M Wehbe, Faraz S Ahmad, Hanyin Wang, and Yuan Luo. 2022. “A Comparative Study of Pretrained Language Models for Long Clinical Text.” Journal of the American Medical Informatics Association 30 (November). https://doi.org/10.1093/jamia/ocac225.
Li, Yilin, Jiaqi Zhu, Congcong Zhang, Yi Yang, Jiawen Zhang, Ying Qiao, and Hongan Wang. 2023. “THGNN: An Embedding-Based Model for Anomaly Detection in Dynamic Heterogeneous Social Networks.” Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3583780.3615079.
Li, Yinghao Aaron, Ali Zare, and Nima Mesgarani. 2021. “StarGANv2-VC: A Diverse, Unsupervised, Non-Parallel Framework for Natural-Sounding Voice Conversion.” arXiv. https://doi.org/10.48550/ARXIV.2107.10394.
Li, Yiwen, Yunguan Fu, Iani Gayo, Qianye Yang, Zhe Min, Shaheer Saeed, Wen Yan, et al. 2022. “Prototypical Few-Shot Segmentation for Cross-Institution Male Pelvic Structures with Spatial Registration.” arXiv. https://doi.org/10.48550/ARXIV.2209.05160.
Li, Youjia, Jianjun Shi, and Zheng Zhang. 2023. “A Novel Approach for Rapid Development Based on ChatGPT and Prompt Engineering.” arXiv. https://doi.org/10.48550/ARXIV.2312.13115.
Li, Yujia, Daniel Tarlow, Marc Brockschmidt, and Richard Zemel. 2015. “Gated Graph Sequence Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.05493.
Li, Zeyu, Jingsheng Gao, Tong Yu, Suncheng Xiang, Jiacheng Ruan, Ting Liu, and Yuzhuo Fu. 2024. “CLAPP: Contrastive Language-Audio Pre-Training in Passive Underwater Vessel Classification.” arXiv. https://doi.org/10.48550/ARXIV.2401.02099.
Li, Zhe, Xiaoyu Wang, Xutao Lv, and Tianbao Yang. 2017. “SEP-Nets: Small and Effective Pattern Networks.” arXiv. https://doi.org/10.48550/ARXIV.1706.03912.
Li, Zhexin, Tong Yang, Peisong Wang, and Jian Cheng. 2022. “Q-ViT: Fully Differentiable Quantization for Vision Transformer.” arXiv. https://doi.org/10.48550/ARXIV.2201.07703.
Li, Zhihan, Youjian Zhao, Jiaqi Han, Ya Su, Rui Jiao, Xidao Wen, and Dan Pei. 2021. “Multivariate Time Series Anomaly Detection and Interpretation Using Hierarchical Inter-Metric and Temporal Embedding.” Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3447548.3467075.
Lialin, Vladislav, Vijeta Deshpande, and Anna Rumshisky. 2023. “Scaling down to Scale up: A Guide to Parameter-Efficient Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2303.15647.
Liang, Chen, Ziqi Liu, Bin Liu, Jun Zhou, Xiaolong Li, Shuang Yang, and Yuan Qi. 2019. “Uncovering Insurance Fraud Conspiracy with Network Learning.” Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3331184.3331372.
Liang, Chen, Wenguan Wang, Tianfei Zhou, Jiaxu Miao, Yawei Luo, and Yi Yang. 2023. “Local-Global Context Aware Transformer for Language-Guided Video Segmentation.” IEEE Transactions on Pattern Analysis and Machine Intelligence 45 (August). https://doi.org/10.1109/tpami.2023.3262578.
Liang, Jiongqian, Peter Jacobs, Jiankai Sun, and Srinivasan Parthasarathy. 2017. “Semi-Supervised Embedding in Attributed Networks with Outliers.” arXiv. https://doi.org/10.48550/ARXIV.1703.08100.
Liang, Shuang, Shouyi Yin, Leibo Liu, Wayne Luk, and Shaojun Wei. 2018. “FP-BNN: Binarized Neural Network on FPGA.” Neurocomputing 275 (January). https://doi.org/10.1016/j.neucom.2017.09.046.
Liang, Yaobo, Chenfei Wu, Ting Song, Wenshan Wu, Yan Xia, Yu Liu, Yang Ou, et al. 2023. “TaskMatrix.AI: Completing Tasks by Connecting Foundation Models with Millions of APIs.” arXiv. https://doi.org/10.48550/ARXIV.2303.16434.
Liang, Yuxuan, Songyu Ke, Junbo Zhang, Xiuwen Yi, and Yu Zheng. 2018. “GeoMAN: Multi-Level Attention Networks for Geo-Sensory Time Series Prediction.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/476.
Liang, Yuxuan, Kun Ouyang, Lin Jing, Sijie Ruan, Ye Liu, Junbo Zhang, David S. Rosenblum, and Yu Zheng. 2019. “UrbanFM.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330646.
Liao, Binbing, Jingqing Zhang, Chao Wu, Douglas McIlwraith, Tong Chen, Shengwen Yang, Yike Guo, and Fei Wu. 2018. “Deep Sequence Learning with Auxiliary Information for Traffic Prediction.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3219895.
Liao, Feng-Ting, Yung-Chieh Chan, Yi-Chang Chen, Chan-Jan Hsu, and Da-shan Shiu. 2023. “Zero-Shot Domain-Sensitive Speech Recognition with Prompt-Conditioning Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2307.10274.
Liao, Hairen, Lingxiao Peng, Zhenchuan Liu, and Xuehua Shen. 2014. “iPinYou Global RTB Bidding Algorithm Competition Dataset.” Proceedings of the Eighth International Workshop on Data Mining for Online Advertising, August. https://doi.org/10.1145/2648584.2648590.
Liao, Jing, Yuan Yao, Lu Yuan, Gang Hua, and Sing Bing Kang. 2017. “Visual Attribute Transfer Through Deep Image Analogy.” arXiv. https://doi.org/10.48550/ARXIV.1705.01088.
Liao, Wenxiong, Zhengliang Liu, Haixing Dai, Shaochen Xu, Zihao Wu, Yiyang Zhang, Xiaoke Huang, et al. 2023. “Differentiating ChatGPT-Generated and Human-Written Medical Texts: Quantitative Study.” JMIR Medical Education 9 (December). https://doi.org/10.2196/48904.
Liberis, Edgar, Łukasz Dudziak, and Nicholas D. Lane. 2021. “μNAS.” Proceedings of the 1st Workshop on Machine Learning and Systems, April. https://doi.org/10.1145/3437984.3458836.
Liebl, Dominik. 2013. “Modeling and Forecasting Electricity Spot Prices: A Functional Data Perspective.” The Annals of Applied Statistics 7 (September). https://doi.org/10.1214/13-aoas652.
Lillicrap, Timothy P., Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. 2015. “Continuous Control with Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1509.02971.
Lim, Jaechang, Seongok Ryu, Jin Woo Kim, and Woo Youn Kim. 2018. “Molecular Generative Model Based on Conditional Variational Autoencoder for de Novo Molecular Design.” arXiv. https://doi.org/10.48550/ARXIV.1806.05805.
Lim, Michael, and Trevor Hastie. 2015. “Learning Interactions via Hierarchical Group-Lasso Regularization.” Journal of Computational and Graphical Statistics 24 (July). https://doi.org/10.1080/10618600.2014.938812.
Lim, Wei Yang Bryan, Nguyen Cong Luong, Dinh Thai Hoang, Yutao Jiao, Ying-Chang Liang, Qiang Yang, Dusit Niyato, and Chunyan Miao. 2019. “Federated Learning in Mobile Edge Networks: A Comprehensive Survey.” arXiv. https://doi.org/10.48550/ARXIV.1909.11875.
Limoyo, Oliver, Trevor Ablett, and Jonathan Kelly. 2022. “Learning Sequential Latent Variable Models from Multimodal Time Series Data.” arXiv. https://doi.org/10.48550/ARXIV.2204.10419.
Lin, Bang, Xiuchong Wang, Yu Dong, Chengfu Huo, Weijun Ren, and Chuanyu Xu. 2021. “Metapaths Guided Neighbors Aggregated Network for?heterogeneous Graph Reasoning.” arXiv. https://doi.org/10.48550/ARXIV.2103.06474.
Lin, Chris, Gerald J. Sun, Krishna C. Bulusu, Jonathan R. Dry, and Marylens Hernandez. 2020. “Graph Neural Networks Including Sparse Interpretability.” arXiv. https://doi.org/10.48550/ARXIV.2007.00119.
Lin, Feng, Hanling Yi, Hongbin Li, Yifan Yang, Xiaotian Yu, Guangming Lu, and Rong Xiao. 2024. “BiTA: Bi-Directional Tuning for Lossless Acceleration in Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.12522.
Lin, Haozhe, Yushun Fan, Jia Zhang, and Bing Bai. 2021. “REST: Reciprocal Framework for Spatiotemporal-Coupled Predictions.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3449928.
Lin, Jeng-Hau, Tianwei Xing, Ritchie Zhao, Zhiru Zhang, Mani Srivastava, Zhuowen Tu, and Rajesh K. Gupta. 2017. “Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration.” arXiv. https://doi.org/10.48550/ARXIV.1707.04693.
Lin, Jiaju, Qin Chen, Jie Zhou, Jian Jin, and Liang He. 2022. “CUP: Curriculum Learning Based Prompt Tuning for Implicit Event Argument Extraction.” arXiv. https://doi.org/10.48550/ARXIV.2205.00498.
Lin, Min, Qiang Chen, and Shuicheng Yan. 2013. “Network in Network.” arXiv. https://doi.org/10.48550/ARXIV.1312.4400.
Lin, Shaohui, Rongrong Ji, Chenqian Yan, Baochang Zhang, Liujuan Cao, Qixiang Ye, Feiyue Huang, and David Doermann. 2019. “Towards Optimal Structured CNN Pruning via Generative Adversarial Learning.” arXiv. https://doi.org/10.48550/ARXIV.1903.09291.
Lin, Tiancheng, Zhimiao Yu, Hongyu Hu, Yi Xu, and Chang Wen Chen. 2023. “Interventional Bag Multi-Instance Learning on Whole-Slide Pathological Images.” arXiv. https://doi.org/10.48550/ARXIV.2303.06873.
Lin, Tsung-Yi, Michael Maire, Serge Belongie, Lubomir Bourdev, Ross Girshick, James Hays, Pietro Perona, Deva Ramanan, C. Lawrence Zitnick, and Piotr Dollár. 2014. “Microsoft COCO: Common Objects in Context.” arXiv. https://doi.org/10.48550/ARXIV.1405.0312.
Lin, Xiaomin, Peter A. Beling, and Randy Cogill. 2014. “Comparison of Multi-Agent and Single-Agent Inverse Learning on a Simulated Soccer Example.” arXiv. https://doi.org/10.48550/ARXIV.1403.6822.
Lin, Yujun, Song Han, Huizi Mao, Yu Wang, and William J. Dally. 2017. “Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training.” arXiv. https://doi.org/10.48550/ARXIV.1712.01887.
Lin, Zeming, Halil Akin, Roshan Rao, Brian Hie, Zhongkai Zhu, Wenting Lu, Nikita Smetanin, et al. 2022. “Evolutionary-Scale Prediction of Atomic Level Protein Structure with a Language Model,” July. https://doi.org/10.1101/2022.07.20.500902.
Lin, Zhouhan, Matthieu Courbariaux, Roland Memisevic, and Yoshua Bengio. 2015. “Neural Networks with Few Multiplications.” arXiv. https://doi.org/10.48550/ARXIV.1510.03009.
Lin, Zhouhan, Minwei Feng, Cicero Nogueira dos Santos, Mo Yu, Bing Xiang, Bowen Zhou, and Yoshua Bengio. 2017. “A Structured Self-Attentive Sentence Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1703.03130.
Linardi, Michele, Yan Zhu, Themis Palpanas, and Eamonn Keogh. 2018. “Matrix Profile x.” Proceedings of the 2018 International Conference on Management of Data, May. https://doi.org/10.1145/3183713.3183744.
Linden, G., B. Smith, and J. York. 2003. “Amazon.com Recommendations: Item-to-Item Collaborative Filtering.” IEEE Internet Computing 7 (January). https://doi.org/10.1109/mic.2003.1167344.
Ling, Wang, Dani Yogatama, Chris Dyer, and Phil Blunsom. 2017. “Program Induction by Rationale Generation : Learning to Solve and Explain Algebraic Word Problems.” arXiv. https://doi.org/10.48550/ARXIV.1705.04146.
Lipton, Zachary C., John Berkowitz, and Charles Elkan. 2015. “A Critical Review of Recurrent Neural Networks for Sequence Learning.” arXiv. https://doi.org/10.48550/ARXIV.1506.00019.
Lipton, Zachary C., David C. Kale, Charles Elkan, and Randall Wetzel. 2015. “Learning to Diagnose with LSTM Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.03677.
Lipton, Zachary C., and Jacob Steinhardt. 2018. “Troubling Trends in Machine Learning Scholarship.” arXiv. https://doi.org/10.48550/ARXIV.1807.03341.
Lisbona, Miguel Angel Alcobendas, Sheide Chammas, and Kuang-chih Lee. 2016. “Optimal Reserve Prices in Upstream Auctions.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939877.
Liu, Aiwei, Xuming Hu, Lijie Wen, and Philip S. Yu. 2023. “A Comprehensive Evaluation of ChatGPT’s Zero-Shot Text-to-SQL Capability.” arXiv. https://doi.org/10.48550/ARXIV.2303.13547.
Liu, Andy T., Po-chun Hsu, and Hung-Yi Lee. 2019. “Unsupervised End-to-End Learning of Discrete Linguistic Units for Voice Conversion.” Interspeech 2019, September. https://doi.org/10.21437/interspeech.2019-2048.
Liu, Bing, and Ian Lane. 2016. “Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling.” arXiv. https://doi.org/10.48550/ARXIV.1609.01454.
Liu, Bing, Harrisen Scells, Guido Zuccon, Wen Hua, and Genghong Zhao. 2021. “ActiveEA: Active Learning for Neural Entity Alignment.” arXiv. https://doi.org/10.48550/ARXIV.2110.06474.
Liu, Bo, Yuqian Jiang, Xiaohan Zhang, Qiang Liu, Shiqi Zhang, Joydeep Biswas, and Peter Stone. 2023. “LLM+p: Empowering Large Language Models with Optimal Planning Proficiency.” arXiv. https://doi.org/10.48550/ARXIV.2304.11477.
Liu, Chengliang, Jie Wen, Xiaoling Luo, and Yong Xu. 2023. “Incomplete Multi-View Multi-Label Learning via Label-Guided Masked View- and Category-Aware Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2303.07180.
Liu, Chia-Wei, Ryan Lowe, Iulian V. Serban, Michael Noseworthy, Laurent Charlin, and Joelle Pineau. 2016. “How NOT to Evaluate Your Dialogue System: An Empirical Study of Unsupervised Evaluation Metrics for Dialogue Response Generation.” arXiv. https://doi.org/10.48550/ARXIV.1603.08023.
Liu, Dairui, Boming Yang, Honghui Du, Derek Greene, Aonghus Lawlor, Ruihai Dong, and Irene Li. 2023. “RecPrompt: A Prompt Tuning Framework for News Recommendation Using Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2312.10463.
Liu, Fuqiang, and Chenchen Liu. 2018. “Towards Accurate and High-Speed Spiking Neuromorphic Systems with Data Quantization-Aware Deep Networks.” Proceedings of the 55th Annual Design Automation Conference, June. https://doi.org/10.1145/3195970.3196131.
Liu, Haitao, Yew-Soon Ong, Xiaobo Shen, and Jianfei Cai. 2018. “When Gaussian Process Meets Big Data: A Review of Scalable GPs.” arXiv. https://doi.org/10.48550/ARXIV.1807.01065.
Liu, Hanmeng, Ruoxi Ning, Zhiyang Teng, Jian Liu, Qiji Zhou, and Yue Zhang. 2023. “Evaluating the Logical Reasoning Ability of ChatGPT and GPT-4.” arXiv. https://doi.org/10.48550/ARXIV.2304.03439.
Liu, Hanxiao, Yuexin Wu, and Yiming Yang. 2017. “Analogical Inference for Multi-Relational Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1705.02426.
Liu, Haohe, Zehua Chen, Yi Yuan, Xinhao Mei, Xubo Liu, Danilo Mandic, Wenwu Wang, and Mark D. Plumbley. 2023. “AudioLDM: Text-to-Audio Generation with Latent Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.12503.
Liu, Haolin, Yujian Zheng, Guanying Chen, Shuguang Cui, and Xiaoguang Han. 2022. “Towards High-Fidelity Single-View Holistic Reconstruction of Indoor Scenes.” arXiv. https://doi.org/10.48550/ARXIV.2207.08656.
Liu, Haotian, Chunyuan Li, Yuheng Li, and Yong Jae Lee. 2023. “Improved Baselines with Visual Instruction Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2310.03744.
Liu, Haotian, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. 2023. “Visual Instruction Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2304.08485.
Liu, Haoxin, Wenli Zhang, Jiaheng Xie, Buomsoo Kim, Zhu Zhang, and Yidong Chai. 2024. “Few-Shot Learning for Chronic Disease Management: Leveraging Large Language Models and Multi-Prompt Engineering with Medical Knowledge Injection.” arXiv. https://doi.org/10.48550/ARXIV.2401.12988.
Liu, Hong, Yucheng Cai, Yuan Zhou, Zhijian Ou, Yi Huang, and Junlan Feng. 2023. “Prompt Pool Based Class-Incremental Continual Learning for Dialog State Tracking.” arXiv. https://doi.org/10.48550/ARXIV.2311.10271.
Liu, Jiawei, Chunqiu Steven Xia, Yuyao Wang, and Lingming Zhang. 2023. “Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation.” arXiv. https://doi.org/10.48550/ARXIV.2305.01210.
Liu, Jiaxi, Yidong Zhang, Xiaoqing Wang, Yuming Deng, and Xingyu Wu. 2019. “Dynamic Pricing on e-Commerce Platform with Deep Reinforcement Learning: A Field Experiment.” arXiv. https://doi.org/10.48550/ARXIV.1912.02572.
Liu, Jiaying, Feng Xia, Xu Feng, Jing Ren, and Huan Liu. 2022. “Deep Graph Learning for Anomalous Citation Detection.” arXiv. https://doi.org/10.48550/ARXIV.2202.11360.
Liu, Jinglin, Chengxi Li, Yi Ren, Feiyang Chen, and Zhou Zhao. 2021. “DiffSinger: Singing Voice Synthesis via Shallow Diffusion Mechanism.” arXiv. https://doi.org/10.48550/ARXIV.2105.02446.
Liu, Jiyuan, Bingyi Lu, Mingkang Xiong, Tao Zhang, and Huilin Xiong. 2023. “Adversarial Attack with Raindrops.” arXiv. https://doi.org/10.48550/ARXIV.2302.14267.
Liu, Kailong, Xiaosong Hu, Zhongbao Wei, Yi Li, and Yan Jiang. 2021. “Modified Gaussian Process Regression Models for Cyclic Capacity Prediction of Lithium-Ion Batteries.” arXiv. https://doi.org/10.48550/ARXIV.2101.00035.
Liu, Kay, Yingtong Dou, Yue Zhao, Xueying Ding, Xiyang Hu, Ruitong Zhang, Kaize Ding, et al. 2022. “PyGOD: A Python Library for Graph Outlier Detection.” arXiv. https://doi.org/10.48550/ARXIV.2204.12095.
Liu, Liang, Boshen Zhang, Jiangning Zhang, Wuhao Zhang, Zhenye Gan, Guanzhong Tian, Wenbing Zhu, Yabiao Wang, and Chengjie Wang. 2023. “MixTeacher: Mining Promising Labels with Mixed Scale Teacher for Semi-Supervised Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2303.09061.
Liu, Ming-Yu, Xun Huang, Arun Mallya, Tero Karras, Timo Aila, Jaakko Lehtinen, and Jan Kautz. 2019. “Few-Shot Unsupervised Image-to-Image Translation.” arXiv. https://doi.org/10.48550/ARXIV.1905.01723.
Liu, Ninghao, Donghwa Shin, and Xia Hu. 2017. “Contextual Outlier Interpretation.” arXiv. https://doi.org/10.48550/ARXIV.1711.10589.
Liu, Ninghao, Qiaoyu Tan, Yuening Li, Hongxia Yang, Jingren Zhou, and Xia Hu. 2019. “Is a Single Vector Enough?” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330967.
Liu, Pengfei, Yiming Ren, Jun Tao, and Zhixiang Ren. 2023. “GIT-Mol: A Multi-Modal Large Language Model for Molecular Science with Graph, Image, and Text.” arXiv. https://doi.org/10.48550/ARXIV.2308.06911.
Liu, Penghang, Rupam Acharyya, Robert E. Tillman, Shunya Kimura, Naoki Masuda, and Ahmet Erdem Sarıyüce. 2023. “Temporal Motifs for Financial Networks: A Study on Mercari, JPMC, and Venmo Platforms.” arXiv. https://doi.org/10.48550/ARXIV.2301.07791.
Liu, Peng, Lemei Zhang, and Jon Atle Gulla. 2023. “Pre-Train, Prompt and Recommendation: A Comprehensive Survey of Language Modelling Paradigm Adaptations in Recommender Systems.” arXiv. https://doi.org/10.48550/ARXIV.2302.03735.
Liu, Qiang, Feng Yu, Shu Wu, and Liang Wang. 2015. “A Convolutional Click Prediction Model.” Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, October. https://doi.org/10.1145/2806416.2806603.
Liu, Qi, Matt J. Kusner, and Phil Blunsom. 2020. “A Survey on Contextual Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.2003.07278.
Liu, Qingju, Mark Kenny, Ramin Nilforooshan, and Payam Barnaghi. 2021. “An Intelligent Bed Sensor System for Non-Contact Respiratory Rate Monitoring.” arXiv. https://doi.org/10.48550/ARXIV.2103.13792.
Liu, Shenghua, Bryan Hooi, and Christos Faloutsos. 2017. “HoloScope: Topology-and-Spike Aware Fraud Detection.” arXiv. https://doi.org/10.48550/ARXIV.1705.02505.
———. 2019. “A Contrast Metric for Fraud Detection in Rich Graphs.” IEEE Transactions on Knowledge and Data Engineering 31 (December). https://doi.org/10.1109/tkde.2018.2876531.
Liu, Shengjun, Ningkang Jiang, and Yuanbin Wu. 2020. “Visual Attack and Defense on Text.” arXiv. https://doi.org/10.48550/ARXIV.2008.10356.
Liu, Shiwei, Tianlong Chen, Zhenyu Zhang, Xuxi Chen, Tianjin Huang, Ajay Jaiswal, and Zhangyang Wang. 2023. “Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!” arXiv. https://doi.org/10.48550/ARXIV.2303.02141.
Liu, Siru, Aileen P. Wright, Barron L. Patterson, Jonathan P. Wanderer, Robert W. Turer, Scott D. Nelson, Allison B. McCoy, Dean F. Sittig, and Adam Wright. 2023. “Assessing the Value of ChatGPT for Clinical Decision Support Optimization,” February. https://doi.org/10.1101/2023.02.21.23286254.
Liu, Ting, Yue Hu, Wansen Wu, Youkai Wang, Kai Xu, and Quanjun Yin. 2023. “Prompt-Based Context- and Domain-Aware Pretraining for Vision and Language Navigation.” arXiv. https://doi.org/10.48550/ARXIV.2309.03661.
Liu, Weijun, and Xiaowei Wang. 2019. “Prediction of Functional microRNA Targets by Integrative Modeling of microRNA Binding and Target Expression Data.” Genome Biology 20 (January). https://doi.org/10.1186/s13059-019-1629-z.
Liu, Wei, Andrew Rabinovich, and Alexander C. Berg. 2015. “ParseNet: Looking Wider to See Better.” arXiv. https://doi.org/10.48550/ARXIV.1506.04579.
Liu, Xiaodong, Yelong Shen, Kevin Duh, and Jianfeng Gao. 2017. “Stochastic Answer Networks for Machine Reading Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.1712.03556.
Liu, Xingchao, Mao Ye, Dengyong Zhou, and Qiang Liu. 2020. “Post-Training Quantization with Multiple Points: Mixed Precision Without Mixed Precision.” arXiv. https://doi.org/10.48550/ARXIV.2002.09049.
Liu, Xinhai, Zhizhong Han, Xin Wen, Yu-Shen Liu, and Matthias Zwicker. 2019. “L2G Auto-Encoder.” Proceedings of the 27th ACM International Conference on Multimedia, October. https://doi.org/10.1145/3343031.3350960.
Liu, Xinyang, Dongsheng Wang, Miaoge Li, Zhibin Duan, Yishi Xu, Bo Chen, and Mingyuan Zhou. 2023. “Patch-Token Aligned Bayesian Prompt Learning for Vision-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.09100.
Liu, Xuan, Xiaoguang Wang, and Stan Matwin. 2018. “Improving the Interpretability of Deep Neural Networks with Knowledge Distillation.” arXiv. https://doi.org/10.48550/ARXIV.1812.10924.
Liu, Yang, Xiang Ao, Zidi Qin, Jianfeng Chi, Jinghua Feng, Hao Yang, and Qing He. 2021. “Pick and Choose: A GNN-Based Imbalanced Learning Approach for Fraud Detection.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3449989.
Liu, Yang, Xiang Ao, Qiwei Zhong, Jinghua Feng, Jiayu Tang, and Qing He. 2020. “Alike and Unlike.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412111.
Liu, Yanqing, Zhihang Xu, Gang Wang, Kuan Chen, Bohan Li, Xu Tan, Jinzhu Li, Lei He, and Sheng Zhao. 2021. “DelightfulTTS: The Microsoft Speech Synthesis System for Blizzard Challenge 2021.” arXiv. https://doi.org/10.48550/ARXIV.2110.12612.
Liu, Yezheng, Zhe Li, Chong Zhou, Yuanchun Jiang, Jianshan Sun, Meng Wang, and Xiangnan He. 2019. “Generative Adversarial Active Learning for Unsupervised Outlier Detection.” IEEE Transactions on Knowledge and Data Engineering. https://doi.org/10.1109/tkde.2019.2905606.
Liu, Yi, Gelei Deng, Zhengzi Xu, Yuekang Li, Yaowen Zheng, Ying Zhang, Lida Zhao, Tianwei Zhang, and Yang Liu. 2023. “Jailbreaking ChatGPT via Prompt Engineering: An Empirical Study.” arXiv. https://doi.org/10.48550/ARXIV.2305.13860.
Liu, Yilun, Shimin Tao, Weibin Meng, Jingyu Wang, Wenbing Ma, Yanqing Zhao, Yuhang Chen, Hao Yang, Yanfei Jiang, and Xun Chen. 2023. “Interpretable Online Log Analysis Using Large Language Models with Prompt Strategies.” arXiv. https://doi.org/10.48550/ARXIV.2308.07610.
Liu, Yinhan, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. “RoBERTa: A Robustly Optimized BERT Pretraining Approach.” arXiv. https://doi.org/10.48550/ARXIV.1907.11692.
Liu, Yinqiu, Hongyang Du, Dusit Niyato, Jiawen Kang, Shuguang Cui, Xuemin Shen, and Ping Zhang. 2023. “Optimizing Mobile-Edge AI-Generated Everything (AIGX) Services by Prompt Engineering: Fundamental, Framework, and Case Study.” arXiv. https://doi.org/10.48550/ARXIV.2309.01065.
Liu, Yisi, Peter Wu, Alan W Black, and Gopala K. Anumanchipalli. 2022. “A Fast and Accurate Pitch Estimation Algorithm Based on the Pseudo Wigner-Ville Distribution.” arXiv. https://doi.org/10.48550/ARXIV.2210.15272.
Liu, Yozen, Xiaolin Shi, Lucas Pierce, and Xiang Ren. 2019. “Characterizing and Forecasting User Engagement with in-App Action Graph: A Case Study of Snapchat.” arXiv. https://doi.org/10.48550/ARXIV.1906.00355.
Liu, Yuan, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, et al. 2023. “MMBench: Is Your Multi-Modal Model an All-Around Player?” arXiv. https://doi.org/10.48550/ARXIV.2307.06281.
Liu, Yuang, Wei Zhang, Jun Wang, and Jianyong Wang. 2021. “Data-Free Knowledge Transfer: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2112.15278.
Liu, Yuanwei, Zhijin Qin, Yunlong Cai, Yue Gao, Geoffrey Ye Li, and Arumugam Nallanathan. 2018. “UAV Communications Based on Non-Orthogonal Multiple Access.” arXiv. https://doi.org/10.48550/ARXIV.1809.05767.
Liu, Yuzhang, Peng Wang, Yingtai Li, Yizhan Shao, and Zhongkai Xu. 2020. “AprilE: Attention with Pseudo Residual Connection for Knowledge Graph Embedding.” Proceedings of the 28th International Conference on Computational Linguistics. https://doi.org/10.18653/v1/2020.coling-main.44.
Liu, Zhenghao, Chenyan Xiong, Maosong Sun, and Zhiyuan Liu. 2019. “Fine-Grained Fact Verification with Kernel Graph Attention Network.” arXiv. https://doi.org/10.48550/ARXIV.1910.09796.
Liu, Zhengliang, Yue Huang, Xiaowei Yu, Lu Zhang, Zihao Wu, Chao Cao, Haixing Dai, et al. 2023. “DeID-GPT: Zero-Shot Medical Text de-Identification by GPT-4.” arXiv. https://doi.org/10.48550/ARXIV.2303.11032.
Liu, Zhiwei, Yingtong Dou, Philip S. Yu, Yutong Deng, and Hao Peng. 2020. “Alleviating the Inconsistency Problem of Applying Graph Neural Network to Fraud Detection.” Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3397271.3401253.
Liu, Zhuang, Jianguo Li, Zhiqiang Shen, Gao Huang, Shoumeng Yan, and Changshui Zhang. 2017. “Learning Efficient Convolutional Networks Through Network Slimming.” arXiv. https://doi.org/10.48550/ARXIV.1708.06519.
Liu, Ziqi, Chaochao Chen, Xinxing Yang, Jun Zhou, Xiaolong Li, and Le Song. 2020. “Heterogeneous Graph Neural Networks for Malicious Account Detection.” arXiv. https://doi.org/10.48550/ARXIV.2002.12307.
Livne, Dor, and Kobi Cohen. 2020. “PoPS: Policy Pruning and Shrinking for Deep Reinforcement Learning.” IEEE Journal of Selected Topics in Signal Processing 14 (May). https://doi.org/10.1109/jstsp.2020.2967566.
Livni, Roi, Daniel Carmon, and Amir Globerson. 2016. “Learning Infinite-Layer Networks: Without the Kernel Trick.” arXiv. https://doi.org/10.48550/ARXIV.1606.05316.
Llinares-López, Felipe, Quentin Berthet, Mathieu Blondel, Olivier Teboul, and Jean-Philippe Vert. 2021. “Deep Embedding and Alignment of Protein Sequences,” November. https://doi.org/10.1101/2021.11.15.468653.
Lloyd, James Robert, David Duvenaud, Roger Grosse, Joshua B. Tenenbaum, and Zoubin Ghahramani. 2014. “Automatic Construction and Natural-Language Description of Nonparametric Regression Models.” arXiv. https://doi.org/10.48550/ARXIV.1402.4304.
Lloyd, Seth, Masoud Mohseni, and Patrick Rebentrost. 2013. “Quantum Algorithms for Supervised and Unsupervised Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1307.0411.
Loh, Wei-Yin. 2009. “Improving the Precision of Classification Trees.” The Annals of Applied Statistics 3 (December). https://doi.org/10.1214/09-aoas260.
Loh, Wei-Yin, and Wei Zheng. 2013. “Regression Trees for Longitudinal and Multiresponse Data.” The Annals of Applied Statistics 7 (March). https://doi.org/10.1214/12-aoas596.
Lohrenz, Timo, Zhengyang Li, and Tim Fingscheidt. 2021. “Multi-Encoder Learning and Stream Fusion for Transformer-Based End-to-End Automatic Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2104.00120.
Lokegaonkar, Vaibhavi, Vijay Jaisankar, Pon Deepika, Madhav Rao, T K Srikanth, Sarbani Mallick, and Manjit Sodhi. 2023. “Introducing SSBD+ Dataset with a Convolutional Pipeline for Detecting Self-Stimulatory Behaviours in Children Using Raw Videos.” arXiv. https://doi.org/10.48550/ARXIV.2311.15072.
Lones, Michael A. 2021. “How to Avoid Machine Learning Pitfalls: A Guide for Academic Researchers.” arXiv. https://doi.org/10.48550/ARXIV.2108.02497.
Longpre, Shayne, Le Hou, Tu Vu, Albert Webson, Hyung Won Chung, Yi Tay, Denny Zhou, et al. 2023. “The Flan Collection: Designing Data and Methods for Effective Instruction Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2301.13688.
Löning, Markus, and Franz Király. 2020. “Forecasting with Sktime: Designing Sktime’s New Forecasting API and Applying It to Replicate and Extend the M4 Study.” arXiv. https://doi.org/10.48550/ARXIV.2005.08067.
Lopes, Raphael Gontijo, Stefano Fenu, and Thad Starner. 2017. “Data-Free Knowledge Distillation for Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1710.07535.
Lopez-Lira, Alejandro, and Yuehua Tang. 2023. “Can ChatGPT Forecast Stock Price Movements? Return Predictability and Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.07619.
Lopez-Paz, David, Léon Bottou, Bernhard Schölkopf, and Vladimir Vapnik. 2015. “Unifying Distillation and Privileged Information.” arXiv. https://doi.org/10.48550/ARXIV.1511.03643.
Lorena, Ana C., Aron I. Maciel, Péricles B. C. de Miranda, Ivan G. Costa, and Ricardo B. C. Prudêncio. 2017. “Data Complexity Meta-Features for Regression Problems.” Machine Learning 107 (December). https://doi.org/10.1007/s10994-017-5681-1.
Lorenzo, Paolo Di, and Gesualdo Scutari. 2016. “NEXT: In-Network Nonconvex Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1602.00591.
Lorenzo-Trueba, Jaime, Thomas Drugman, Javier Latorre, Thomas Merritt, Bartosz Putrycz, Roberto Barra-Chicote, Alexis Moinet, and Vatsal Aggarwal. 2019. “Towards Achieving Robust Universal Neural Vocoding.” Interspeech 2019, September. https://doi.org/10.21437/interspeech.2019-1424.
Lou, Renze, Kai Zhang, and Wenpeng Yin. 2023. “A Comprehensive Survey on Instruction Following.” arXiv. https://doi.org/10.48550/ARXIV.2303.10475.
Lou, Yin, and Mikhail Obukhov. 2017. “BDT.” Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3097983.3098175.
Louizos, Christos, Karen Ullrich, and Max Welling. 2017. “Bayesian Compression for Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1705.08665.
Low, Yucheng, Joseph E. Gonzalez, Aapo Kyrola, Danny Bickson, Carlos E. Guestrin, and Joseph Hellerstein. 2014. “GraphLab: A New Framework for Parallel Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1408.2041.
Lowe, Gavin. 1995. “An Attack on the Needham-Schroeder Public-Key Authentication Protocol.” Information Processing Letters 56 (November). https://doi.org/10.1016/0020-0190(95)00144-2.
Lowe, Ryan, Yi Wu, Aviv Tamar, Jean Harb, Pieter Abbeel, and Igor Mordatch. 2017. “Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments.” arXiv. https://doi.org/10.48550/ARXIV.1706.02275.
Lu, Chaochao, and Xiaoou Tang. 2014. “Surpassing Human-Level Face Verification Performance on LFW with GaussianFace.” arXiv. https://doi.org/10.48550/ARXIV.1404.3840.
Lu, Haolan, and Bradley P. Carlin. 2005. “Bayesian Areal Wombling for Geographical Boundary Analysis.” Geographical Analysis 37 (July). https://doi.org/10.1111/j.1538-4632.2005.00624.x.
Lu, Hui, Zhiyong Wu, Xixin Wu, Xu Li, Shiyin Kang, Xunying Liu, and Helen Meng. 2021. “VAENAR-TTS: Variational Auto-Encoder Based Non-AutoRegressive Text-to-Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.2107.03298.
Lu, Liang, Michelle Guo, and Steve Renals. 2016. “Knowledge Distillation for Small-Footprint Highway Networks.” arXiv. https://doi.org/10.48550/ARXIV.1608.00892.
Lu, Liang, Naoyuki Kanda, Jinyu Li, and Yifan Gong. 2021. “Streaming End-to-End Multi-Talker Speech Recognition.” IEEE Signal Processing Letters 28. https://doi.org/10.1109/lsp.2021.3070817.
Lu, Liang, Lingpeng Kong, Chris Dyer, and Noah A. Smith. 2017. “Multitask Learning with CTC and Segmental CRF for Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1702.06378.
Lu, Lu, Xuhui Meng, Zhiping Mao, and George Em Karniadakis. 2021. “DeepXDE: A Deep Learning Library for Solving Differential Equations.” SIAM Review 63 (January). https://doi.org/10.1137/19m1274067.
Lu, Ming Y., Drew F. K. Williamson, Tiffany Y. Chen, Richard J. Chen, Matteo Barbieri, and Faisal Mahmood. 2020. “Data Efficient and Weakly Supervised Computational Pathology on Whole Slide Images.” arXiv. https://doi.org/10.48550/ARXIV.2004.09666.
Lu, Pan, Baolin Peng, Hao Cheng, Michel Galley, Kai-Wei Chang, Ying Nian Wu, Song-Chun Zhu, and Jianfeng Gao. 2023. “Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.09842.
Lu, Qingyu, Baopu Qiu, Liang Ding, Kanjian Zhang, Tom Kocmi, and Dacheng Tao. 2023. “Error Analysis Prompting Enables Human-Like Translation Evaluation in Large Language Models: A Case Study on ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2303.13809.
Lu, Shuai, Daya Guo, Shuo Ren, Junjie Huang, Alexey Svyatkovskiy, Ambrosio Blanco, Colin Clement, et al. 2021. “CodeXGLUE: A Machine Learning Benchmark Dataset for Code Understanding and Generation.” arXiv. https://doi.org/10.48550/ARXIV.2102.04664.
Lu, Yen-Ju, Zhong-Qiu Wang, Shinji Watanabe, Alexander Richard, Cheng Yu, and Yu Tsao. 2022. “Conditional Diffusion Probabilistic Model for Speech Enhancement.” ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), May. https://doi.org/10.1109/icassp43922.2022.9746901.
Lu, Yi-Ju, and Cheng-Te Li. 2020. “GCAN: Graph-Aware Co-Attention Networks for Explainable Fake News Detection on Social Media.” Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/2020.acl-main.48.
Lu, Yue, Renjie Wu, Abdullah Mueen, Maria A. Zuluaga, and Eamonn Keogh. 2022. “Matrix Profile XXIV: Scaling Time Series Anomaly Detection to Trillions of Datapoints and Ultra-Fast Arriving Data Streams.” Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3534678.3539271.
Lu, Yujie, Shengyu Zhang, Yingxuan Huang, Luyao Wang, Xinyao Yu, Zhou Zhao, and Fei Wu. 2021. “Future-Aware Diverse Trends Framework for Recommendation.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3449791.
Luan, Fujun, Sylvain Paris, Eli Shechtman, and Kavita Bala. 2017. “Deep Photo Style Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1703.07511.
Luan, Yi, Dave Wadden, Luheng He, Amy Shah, Mari Ostendorf, and Hannaneh Hajishirzi. 2019. “A General Framework for Information Extraction Using Dynamic Span Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1904.03296.
Lucchese, Claudio, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Nicola Tonellotto, and Rossano Venturini. 2015. “QuickScorer.” Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, August. https://doi.org/10.1145/2766462.2767733.
Lucchese, Claudio, Franco Maria Nardini, Raffaele Perego, Salvatore Orlando, and Salvatore Trani. 2018. “Selective Gradient Boosting for Effective Learning to Rank.” The 41st International ACM SIGIR Conference on Research &Amp; Development in Information Retrieval, June. https://doi.org/10.1145/3209978.3210048.
Lundberg, Scott M., Gabriel G. Erion, and Su-In Lee. 2018. “Consistent Individualized Feature Attribution for Tree Ensembles.” arXiv. https://doi.org/10.48550/ARXIV.1802.03888.
Luo, Chen, Jian-Guang Lou, Qingwei Lin, Qiang Fu, Rui Ding, Dongmei Zhang, and Zhe Wang. 2014. “Correlating Events with Time Series for Incident Diagnosis.” Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2623330.2623374.
Luo, Chenxu, Xiaodong Yang, and Alan Yuille. 2021. “Exploring Simple 3D Multi-Object Tracking for Autonomous Driving.” arXiv. https://doi.org/10.48550/ARXIV.2108.10312.
Luo, Chunjie, Jianfeng Zhan, Lei Wang, and Qiang Yang. 2017. “Cosine Normalization: Using Cosine Similarity Instead of Dot Product in Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1702.05870.
Luo, Dongsheng, Wei Cheng, Dongkuan Xu, Wenchao Yu, Bo Zong, Haifeng Chen, and Xiang Zhang. 2020. “Parameterized Explainer for Graph Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.2011.04573.
Luo, Jian-Hao, and Jianxin Wu. 2018. “AutoPruner: An End-to-End Trainable Filter Pruning Method for Efficient Deep Model Inference.” arXiv. https://doi.org/10.48550/ARXIV.1805.08941.
Luo, Jian-Hao, Jianxin Wu, and Weiyao Lin. 2017. “ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression.” arXiv. https://doi.org/10.48550/ARXIV.1707.06342.
Luo, Jiaxin, Qingjun Meng, and Yan Cai. 2018. “Analysis of the Impact of Artificial Intelligence Application on the Development of Accounting Industry.” Open Journal of Business and Management 06. https://doi.org/10.4236/ojbm.2018.64063.
Luo, Linkai, and Yue Wang. 2019. “EmotionX-HSU: Adopting Pre-Trained BERT for Emotion Classification.” arXiv. https://doi.org/10.48550/ARXIV.1907.09669.
Luo, Yun, Zhen Yang, Fandong Meng, Yafu Li, Jie Zhou, and Yue Zhang. 2023. “An Empirical Study of Catastrophic Forgetting in Large Language Models During Continual Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2308.08747.
Luong, Minh-Thang, Quoc V. Le, Ilya Sutskever, Oriol Vinyals, and Lukasz Kaiser. 2015. “Multi-Task Sequence to Sequence Learning.” arXiv. https://doi.org/10.48550/ARXIV.1511.06114.
Luong, Minh-Thang, and Christopher D. Manning. 2016. “Achieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models.” arXiv. https://doi.org/10.48550/ARXIV.1604.00788.
Luong, Minh-Thang, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, and Wojciech Zaremba. 2014. “Addressing the Rare Word Problem in Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1410.8206.
Luong, Thang, Hieu Pham, and Christopher D. Manning. 2015. “Effective Approaches to Attention-Based Neural Machine Translation.” Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/d15-1166.
Luxburg, Ulrike von. 2007. “A Tutorial on Spectral Clustering.” arXiv. https://doi.org/10.48550/ARXIV.0711.0189.
Lv, Fuyu, Taiwei Jin, Changlong Yu, Fei Sun, Quan Lin, Keping Yang, and Wilfred Ng. 2019. “SDM.” Proceedings of the 28th ACM International Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3357384.3357818.
Lv, Ge, and Lei Chen. 2023. “On Data-Aware Global Explainability of Graph Neural Networks.” Proceedings of the VLDB Endowment 16 (July). https://doi.org/10.14778/3611479.3611538.
Lv, Ge, Chen Jason Zhang, and Lei Chen. 2023. “HENCE-x: Toward Heterogeneity-Agnostic Multi-Level Explainability for Deep Graph Networks.” Proceedings of the VLDB Endowment 16 (July). https://doi.org/10.14778/3611479.3611503.
Lybarger, Kevin, Meliha Yetisgen, and Özlem Uzuner. 2023. “The 2022 N2c2/UW Shared Task on Extracting Social Determinants of Health.” Journal of the American Medical Informatics Association 30 (April). https://doi.org/10.1093/jamia/ocad012.
Lym, Sangkug, Esha Choukse, Siavash Zangeneh, Wei Wen, Sujay Sanghavi, and Mattan Erez. 2019. “PruneTrain.” Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, November. https://doi.org/10.1145/3295500.3356156.
Lyu, Chenyang, Jitao Xu, and Longyue Wang. 2023. “New Trends in Machine Translation Using Large Language Models: Case Examples with ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2305.01181.
Lyu, Qing, Shreya Havaldar, Adam Stein, Li Zhang, Delip Rao, Eric Wong, Marianna Apidianaki, and Chris Callison-Burch. 2023. “Faithful Chain-of-Thought Reasoning.” arXiv. https://doi.org/10.48550/ARXIV.2301.13379.
Ma, Chengcheng, Yang Liu, Jiankang Deng, Lingxi Xie, Weiming Dong, and Changsheng Xu. 2022. “Understanding and Mitigating Overfitting in Prompt Tuning for Vision-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2211.02219.
Ma, Chong, Zihao Wu, Jiaqi Wang, Shaochen Xu, Yaonai Wei, Zhengliang Liu, Xi Jiang, et al. 2023. “ImpressionGPT: An Iterative Optimizing Framework for Radiology Report Summarization with ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2304.08448.
Ma, Guixiang, Lifang He, Chun-Ta Lu, Weixiang Shao, Philip S. Yu, Alex D. Leow, and Ann B. Ragin. 2017. “Multi-View Clustering with Graph Embedding for Connectome Analysis.” Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3132847.3132909.
Ma, Haichuan, Dong Liu, Ning Yan, Houqiang Li, and Feng Wu. 2022. “End-to-End Optimized Versatile Image Compression with Wavelet-Like Transform.” IEEE Transactions on Pattern Analysis and Machine Intelligence 44 (March). https://doi.org/10.1109/tpami.2020.3026003.
Ma, Jin, Philip Protter, and Jiongmin Yong. 1994. “Solving Forward-Backward Stochastic Differential Equations Explicitly — a Four Step Scheme.” Probability Theory and Related Fields 98 (September). https://doi.org/10.1007/bf01192258.
Ma, Pingchuan, Alexandros Haliassos, Adriana Fernandez-Lopez, Honglie Chen, Stavros Petridis, and Maja Pantic. 2023. “Auto-AVSR: Audio-Visual Speech Recognition with Automatic Labels.” ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), June. https://doi.org/10.1109/icassp49357.2023.10096889.
Ma, Shiqian, Donald Goldfarb, and Lifeng Chen. 2009. “Fixed Point and Bregman Iterative Methods for Matrix Rank Minimization.” arXiv. https://doi.org/10.48550/ARXIV.0905.1643.
Ma, Xiaoxiao, Jia Wu, Jian Yang, and Quan Z. Sheng. 2023. “Towards Graph-Level Anomaly Detection via Deep Evolutionary Mapping.” Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3580305.3599524.
Ma, Xinyin, Xinchao Wang, Gongfan Fang, Yongliang Shen, and Weiming Lu. 2022. “Prompting to Distill: Boosting Data-Free Knowledge Distillation via Reinforced Prompt.” arXiv. https://doi.org/10.48550/ARXIV.2205.07523.
Ma, Yifei, Balakrishnan (Murali) Narayanaswamy, Haibin Lin, and Hao Ding. 2020. “Temporal-Contextual Recommendation in Real-Time.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403278.
Ma, Yong, Senlin Luo, Yu-Ming Shang, Yifei Zhang, and Zhengjun Li. 2024. “CodePrompt: Improving Source Code-Related Classification with Knowledge Features Through Prompt Learning.” arXiv. https://doi.org/10.48550/ARXIV.2401.05544.
Ma, Yuqing, Xianglong Liu, Shihao Bai, Lei Wang, Dailan He, and Aishan Liu. 2019. “Coarse-to-Fine Image Inpainting via Region-Wise Convolutions and Non-Local Correlation.” Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, August. https://doi.org/10.24963/ijcai.2019/433.
Maaten, Laurens van der, and Geoffrey Hinton. 2011. “Visualizing Non-Metric Similarities in Multiple Maps.” Machine Learning 87 (December). https://doi.org/10.1007/s10994-011-5273-4.
“Machine Learning and Knowledge Discovery in Databases.” 2020. Communications in Computer and Information Science. https://doi.org/10.1007/978-3-030-43823-4.
“Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track.” 2021. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-86514-6.
“Machine Learning: ECML 2005.” 2005. Lecture Notes in Computer Science. https://doi.org/10.1007/11564096.
MacKay, David J. C. 1992a. “Bayesian Methods for Adaptive Models.” https://doi.org/10.7907/H3A1-WM07.
———. 1992b. “A Practical Bayesian Framework for Backpropagation Networks.” Neural Computation 4 (May). https://doi.org/10.1162/neco.1992.4.3.448.
Mackenzie, Joel, Andrew Trotman, and Jimmy Lin. 2022. “Efficient Document-at-a-Time and Score-at-a-Time Query Evaluation for Learned Sparse Representations.” ACM Transactions on Information Systems, December. https://doi.org/10.1145/3576922.
Maclaurin, Dougal, David Duvenaud, and Ryan P. Adams. 2015. “Gradient-Based Hyperparameter Optimization Through Reversible Learning.” arXiv. https://doi.org/10.48550/ARXIV.1502.03492.
Madaan, Aman, Niket Tandon, Peter Clark, and Yiming Yang. 2022. “Memory-Assisted Prompt Editing to Improve GPT-3 After Deployment.” arXiv. https://doi.org/10.48550/ARXIV.2201.06009.
Madaan, Aman, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, et al. 2023. “Self-Refine: Iterative Refinement with Self-Feedback.” arXiv. https://doi.org/10.48550/ARXIV.2303.17651.
Madani, Ali, Ben Krause, Eric R. Greene, Subu Subramanian, Benjamin P. Mohr, James M. Holton, Jose Luis Olmos, et al. 2023. “Large Language Models Generate Functional Protein Sequences Across Diverse Families.” Nature Biotechnology 41 (January). https://doi.org/10.1038/s41587-022-01618-2.
Maddison, Chris J. 2016. “A Poisson Process Model for Monte Carlo.” arXiv. https://doi.org/10.48550/ARXIV.1602.05986.
Madhyastha, Meghana, Gongkai Li, Veronika Strnadová-Neeley, James Browne, Joshua T. Vogelstein, Randal Burns, and Carey E. Priebe. 2020. “Geodesic Forests.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403094.
Madigan, D., E. Einahrawy, R. P. Martin, Wen-Hua Ju, P. Krishnan, and A. S. Krishnakumar. n.d. “Bayesian Indoor Positioning Systems.” Proceedings IEEE 24th Annual Joint Conference of the IEEE Computer and Communications Societies. https://doi.org/10.1109/infcom.2005.1498348.
Madotto, Andrea, Zhaojiang Lin, Genta Indra Winata, and Pascale Fung. 2021. “Few-Shot Bot: Prompt-Based Learning for Dialogue Systems.” arXiv. https://doi.org/10.48550/ARXIV.2110.08118.
Madry, Aleksander, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2017. “Towards Deep Learning Models Resistant to Adversarial Attacks.” arXiv. https://doi.org/10.48550/ARXIV.1706.06083.
Mafla, Andrés, Rafael Sampaio de Rezende, Lluís Gómez, Diane Larlus, and Dimosthenis Karatzas. 2020. “StacMR: Scene-Text Aware Cross-Modal Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.2012.04329.
Magee, Andrew F., Michael D. Karcher, Frederick A. Matsen, and Vladimir N. Minin. 2021. “How Trustworthy Is Your Tree? Bayesian Phylogenetic Effective Sample Size Through the Lens of Monte Carlo Error.” arXiv. https://doi.org/10.48550/ARXIV.2109.07629.
Mahadeokar, Jay, Yuan Shangguan, Duc Le, Gil Keren, Hang Su, Thong Le, Ching-Feng Yeh, Christian Fuegen, and Michael L. Seltzer. 2020. “Alignment Restricted Streaming Recurrent Neural Network Transducer.” arXiv. https://doi.org/10.48550/ARXIV.2011.03072.
Mahbub, Ridwan, Ifrad Khan, Samiha Anuva, Md Shahriar, Md Tahmid Rahman Laskar, and Sabbir Ahmed. 2023. “Unveiling the Essence of Poetry: Introducing a Comprehensive Dataset and Benchmark for Poem Summarization.” Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/2023.emnlp-main.920.
Mahimkar, Ajay, Zihui Ge, Jia Wang, Jennifer Yates, Yin Zhang, Joanne Emmons, Brian Huntley, and Mark Stockert. 2011. “Rapid Detection of Maintenance Induced Changes in Service Performance.” Proceedings of the Seventh COnference on Emerging Networking EXperiments and Technologies, December. https://doi.org/10.1145/2079296.2079309.
Mahmoud, Anas, Jordan S. K. Hu, and Steven L. Waslander. 2022. “Dense Voxel Fusion for 3D Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2203.00871.
Mai, Uyen, and Siavash Mirarab. 2020. “Log Transformation Improves Dating of Phylogenies.” Molecular Biology and Evolution 38 (September). https://doi.org/10.1093/molbev/msaa222.
Maier, Andreas, Christopher Syben, Tobias Lasser, and Christian Riess. 2018. “A Gentle Introduction to Deep Learning in Medical Image Processing.” arXiv. https://doi.org/10.48550/ARXIV.1810.05401.
Mairal, Julien. 2014. “Incremental Majorization-Minimization Optimization with Application to Large-Scale Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1402.4419.
Maji, Subhransu, Esa Rahtu, Juho Kannala, Matthew Blaschko, and Andrea Vedaldi. 2013. “Fine-Grained Visual Classification of Aircraft.” arXiv. https://doi.org/10.48550/ARXIV.1306.5151.
Major, Bence, Daniel Fontijne, Amin Ansari, Ravi Teja Sukhavasi, Radhika Gowaikar, Michael Hamilton, Sean Lee, Slawomir Grzechnik, and Sundar Subramanian. 2019. “Vehicle Detection with Automotive Radar Using Deep Learning on Range-Azimuth-Doppler Tensors.” 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), October. https://doi.org/10.1109/iccvw.2019.00121.
Mäkela, Satu-Marja, Arttu Lämsä, Janne S. Keränen, Jussi Liikka, Jussi Ronkainen, Johannes Peltola, Juha Häikiö, Sari Järvinen, and Miguel Bordallo López. 2021. “Introducing VTT-ConIot: A Realistic Dataset for Activity Recognition of Construction Workers Using IMU Devices.” Sustainability 14 (December). https://doi.org/10.3390/su14010220.
Makoviychuk, Viktor, Lukasz Wawrzyniak, Yunrong Guo, Michelle Lu, Kier Storey, Miles Macklin, David Hoeller, et al. 2021. “Isaac Gym: High Performance GPU-Based Physics Simulation for Robot Learning.” arXiv. https://doi.org/10.48550/ARXIV.2108.10470.
Maksymiuk, Szymon, Alicja Gosiewska, and Przemyslaw Biecek. 2020. “Landscape of r Packages for eXplainable Artificial Intelligence.” arXiv. https://doi.org/10.48550/ARXIV.2009.13248.
Malherbe, Cédric, and Nicolas Vayatis. 2017. “Global Optimization of Lipschitz Functions.” arXiv. https://doi.org/10.48550/ARXIV.1703.02628.
Malhotra, Pankaj, Anusha Ramakrishnan, Gaurangi Anand, Lovekesh Vig, Puneet Agarwal, and Gautam Shroff. 2016. “LSTM-Based Encoder-Decoder for Multi-Sensor Anomaly Detection.” arXiv. https://doi.org/10.48550/ARXIV.1607.00148.
Mallinar, Neil, Abhishek Shah, Rajendra Ugrani, Ayush Gupta, Manikandan Gurusankar, Tin Kam Ho, Q. Vera Liao, et al. 2018. “Bootstrapping Conversational Agents with Weak Supervision.” arXiv. https://doi.org/10.48550/ARXIV.1812.06176.
Manakul, Potsawee, Mark J. F. Gales, and Linlin Wang. 2020. “Abstractive Spoken Document Summarization Using Hierarchical Model with Multi-Stage Attention Diversity Optimization.” Interspeech 2020, October. https://doi.org/10.21437/interspeech.2020-1683.
Manderson, Andrew A., and Robert J. B. Goudie. 2023. “Combining Chains of Bayesian Models with Markov Melding.” Bayesian Analysis 18 (September). https://doi.org/10.1214/22-ba1327.
Mandros, Panagiotis, Mario Boley, and Jilles Vreeken. 2018. “Discovering Reliable Dependencies from Data: Hardness and Improved Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.1809.05467.
Mania, Horia, Xinghao Pan, Dimitris Papailiopoulos, Benjamin Recht, Kannan Ramchandran, and Michael I. Jordan. 2015. “Perturbed Iterate Analysis for Asynchronous Stochastic Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1507.06970.
Manna, Zohar, and Richard J. Waldinger. 1971. “Toward Automatic Program Synthesis.” Communications of the ACM 14 (March). https://doi.org/10.1145/362566.362568.
Mao, Hongzi, Malte Schwarzkopf, Shaileshh Bojja Venkatakrishnan, Zili Meng, and Mohammad Alizadeh. 2018. “Learning Scheduling Algorithms for Data Processing Clusters.” arXiv. https://doi.org/10.48550/ARXIV.1810.01963.
Mao, Huina, Scott Counts, and Johan Bollen. 2011. “Predicting Financial Markets: Comparing Survey, News, Twitter and Search Engine Data.” arXiv. https://doi.org/10.48550/ARXIV.1112.1051.
Mao, Huizi, Song Han, Jeff Pool, Wenshuo Li, Xingyu Liu, Yu Wang, and William J. Dally. 2017. “Exploring the Regularity of Sparse Structure in Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1705.08922.
Mao, Junhua, Jonathan Huang, Alexander Toshev, Oana Camburu, Alan Yuille, and Kevin Murphy. 2015. “Generation and Comprehension of Unambiguous Object Descriptions.” arXiv. https://doi.org/10.48550/ARXIV.1511.02283.
Mao, Xudong, Qing Li, Haoran Xie, Raymond Y. K. Lau, Zhen Wang, and Stephen Paul Smolley. 2016. “Least Squares Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.04076.
———. 2017. “Least Squares Generative Adversarial Networks.” 2017 IEEE International Conference on Computer Vision (ICCV), October. https://doi.org/10.1109/iccv.2017.304.
Marbac, Matthieu, and Mohammed Sedki. 2015. “A Family of Blockwise One-Factor Distributions for Modelling High-Dimensional Binary Data.” arXiv. https://doi.org/10.48550/ARXIV.1511.01343.
Margossian, Charles C., and Andrew Gelman. 2023. “For How Many Iterations Should We Run Markov Chain Monte Carlo?” arXiv. https://doi.org/10.48550/ARXIV.2311.02726.
Mariani, Giovanni, Florian Scheidegger, Roxana Istrate, Costas Bekas, and Cristiano Malossi. 2018. “BAGAN: Data Augmentation with Balancing GAN.” arXiv. https://doi.org/10.48550/ARXIV.1803.09655.
Marinescu, Andrei, Ivana Dusparic, Adam Taylor, Vinny Cahill, and Siobhán Clarke. 2014. “Decentralised Multi-Agent Reinforcement Learning for Dynamic and Uncertain Environments.” arXiv. https://doi.org/10.48550/ARXIV.1409.4561.
Marino, Kenneth, Ruslan Salakhutdinov, and Abhinav Gupta. 2016. “The More You Know: Using Knowledge Graphs for Image Classification.” arXiv. https://doi.org/10.48550/ARXIV.1612.04844.
Marivate, Vukosi, and Tshephisho Sefara. 2020. “Improving Short Text Classification Through Global Augmentation Methods.” Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-030-57321-8_21.
Markov, Igor L., Hanson Wang, Nitya Kasturi, Shaun Singh, Sze Wai Yuen, Mia Garrard, Sarah Tran, et al. 2021. “Looper: An End-to-End ML Platform for Product Decisions.” arXiv. https://doi.org/10.48550/ARXIV.2110.07554.
Marmin, Sébastien, and Maurizio Filippone. 2022. “Deep Gaussian Processes for Calibration of Computer Models (with Discussion).” Bayesian Analysis 17 (December). https://doi.org/10.1214/21-ba1293.
Maroulas, Vasileios, Cassie Putman Micucci, and Farzana Nasrin. 2022. “Bayesian Topological Learning for Classifying the Structure of Biological Networks.” Bayesian Analysis 17 (September). https://doi.org/10.1214/21-ba1270.
Marsden, Robert A., Mario Döbler, and Bin Yang. 2022. “Introducing Intermediate Domains for Effective Self-Training During Test-Time.” arXiv. https://doi.org/10.48550/ARXIV.2208.07736.
Martens, James, and Roger Grosse. 2015. “Optimizing Neural Networks with Kronecker-Factored Approximate Curvature.” arXiv. https://doi.org/10.48550/ARXIV.1503.05671.
Martin, Andrew D., and Kevin M. Quinn. 2002. “Dynamic Ideal Point Estimation via Markov Chain Monte Carlo for the u.s. Supreme Court, 1953–1999.” Political Analysis 10. https://doi.org/10.1093/pan/10.2.134.
Martinez, Julieta, Jashan Shewakramani, Ting Wei Liu, Ioan Andrei Bârsan, Wenyuan Zeng, and Raquel Urtasun. 2020. “Permute, Quantize, and Fine-Tune: Efficient Compression of Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2010.15703.
Martínez-Velasco, Antonieta, Lourdes Martínez-Villaseñor, and Luis Miralles-Pechuán. 2018. “Machine Learning Approach for Pre-Eclampsia Risk Factors Association.” Proceedings of the 4th EAI International Conference on Smart Objects and Technologies for Social Good, November. https://doi.org/10.1145/3284869.3284912.
Masse, Nicolas Y., Gregory D. Grant, and David J. Freedman. 2018. “Alleviating Catastrophic Forgetting Using Context-Dependent Gating and Synaptic Stabilization.” Proceedings of the National Academy of Sciences 115 (October). https://doi.org/10.1073/pnas.1803839115.
Mathys, Christoph. 2011. “A Bayesian Foundation for Individual Learning Under Uncertainty.” Frontiers in Human Neuroscience 5. https://doi.org/10.3389/fnhum.2011.00039.
Maurer, Ueli M. 1995. “Fast Generation of Prime Numbers and Secure Public-Key Cryptographic Parameters.” Journal of Cryptology 8 (September). https://doi.org/10.1007/bf00202269.
Maus, Natalie, Patrick Chao, Eric Wong, and Jacob Gardner. 2023. “Black Box Adversarial Prompting for Foundation Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.04237.
“Maximum Entropy and Bayesian Methods.” 1992. https://doi.org/10.1007/978-94-017-2219-3.
———. 1996. https://doi.org/10.1007/978-94-011-5430-7.
Mazuz, Eyal, Guy Shtar, Bracha Shapira, and Lior Rokach. 2023. “Molecule Generation Using Transformers and Policy Gradient Reinforcement Learning.” Scientific Reports 13 (May). https://doi.org/10.1038/s41598-023-35648-w.
Mazyavkina, Nina, Sergey Sviridov, Sergei Ivanov, and Evgeny Burnaev. 2020. “Reinforcement Learning for Combinatorial Optimization: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2003.03600.
McCann, Bryan, Nitish Shirish Keskar, Caiming Xiong, and Richard Socher. 2018. “The Natural Language Decathlon: Multitask Learning as Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1806.08730.
McFee, Brian, Matt McVicar, Stefan Balke, Vincent Lostanlen, Carl Thomé, Colin Raffel, Dana Lee, et al. 2019. “Librosa/Librosa: 0.6.3,” February. https://doi.org/10.5281/ZENODO.2564164.
McFee, Brian, Matt McVicar, Colin Raffel, and Douglas Repetto. 2014. “Librosa: V0.3.1,” November. https://doi.org/10.5281/ZENODO.12714.
McInerney, James, Benjamin Lacker, Samantha Hansen, Karl Higley, Hugues Bouchard, Alois Gruson, and Rishabh Mehrotra. 2018. “Explore, Exploit, and Explain.” Proceedings of the 12th ACM Conference on Recommender Systems, September. https://doi.org/10.1145/3240323.3240354.
McKee, Forrest, and David Noever. 2023. “Chatbots in a Honeypot World.” arXiv. https://doi.org/10.48550/ARXIV.2301.03771.
McMahan, H. Brendan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Agüera y Arcas. 2016. “Communication-Efficient Learning of Deep Networks from Decentralized Data.” arXiv. https://doi.org/10.48550/ARXIV.1602.05629.
McNamara, Danielle S., Irwin B. Levinstein, and Chutima Boonthum. 2004. “iSTART: Interactive Strategy Training for Active Reading and Thinking.” Behavior Research Methods, Instruments, &Amp; Computers 36 (May). https://doi.org/10.3758/bf03195567.
Megahed, Fadel M., Ying-Ju Chen, Joshua A. Ferris, Sven Knoth, and L. Allison Jones-Farmer. 2023. “How Generative AI Models Such as ChatGPT Can Be (Mis)used in SPC Practice, Education, and Research? An Exploratory Study.” Quality Engineering, June. https://doi.org/10.1080/08982112.2023.2206479.
Mehri, Soroush, Kundan Kumar, Ishaan Gulrajani, Rithesh Kumar, Shubham Jain, Jose Sotelo, Aaron Courville, and Yoshua Bengio. 2016. “SampleRNN: An Unconditional End-to-End Neural Audio Generation Model.” arXiv. https://doi.org/10.48550/ARXIV.1612.07837.
Mehta, Sneha, Huzefa Rangwala, and Naren Ramakrishnan. 2022. “Improving Zero-Shot Event Extraction via Sentence Simplification.” arXiv. https://doi.org/10.48550/ARXIV.2204.02531.
Mei, Xinhao, Chutong Meng, Haohe Liu, Qiuqiang Kong, Tom Ko, Chengqi Zhao, Mark D. Plumbley, Yuexian Zou, and Wenwu Wang. 2023. “WavCaps: A ChatGPT-Assisted Weakly-Labelled Audio Captioning Dataset for Audio-Language Multimodal Research.” arXiv. https://doi.org/10.48550/ARXIV.2303.17395.
Meidani, Mohammadreza, and Behboud Mashoufi. 2016. “Introducing New Algorithms for Realising an FIR Filter with Less Hardware in Order to Eliminate Power Line Interference from the ECG Signal.” IET Signal Processing 10 (September). https://doi.org/10.1049/iet-spr.2015.0552.
Meila, Marina, and David Heckerman. 2013. “An Experimental Comparison of Several Clustering and Initialization Methods.” arXiv. https://doi.org/10.48550/ARXIV.1301.7401.
Meinshausen, Nicolai. 2010. “Node Harvest.” The Annals of Applied Statistics 4 (December). https://doi.org/10.1214/10-aoas367.
Meka, Raghu, Prateek Jain, and Inderjit S. Dhillon. 2009. “Guaranteed Rank Minimization via Singular Value Projection.” arXiv. https://doi.org/10.48550/ARXIV.0909.5457.
Melacci, Stefano, and Mikhail Belkin. 2009. “Laplacian Support Vector Machines Trained in the Primal.” arXiv. https://doi.org/10.48550/ARXIV.0909.5422.
Mellempudi, Naveen, Abhisek Kundu, Dheevatsa Mudigere, Dipankar Das, Bharat Kaul, and Pradeep Dubey. 2017. “Ternary Neural Networks with Fine-Grained Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1705.01462.
Menegola, Afonso, Michel Fornaciali, Ramon Pires, Sandra Avila, and Eduardo Valle. 2016. “Towards Automated Melanoma Screening: Exploring Transfer Learning Schemes.” arXiv. https://doi.org/10.48550/ARXIV.1609.01228.
Meng, Rui, Sanqiang Zhao, Shuguang Han, Daqing He, Peter Brusilovsky, and Yu Chi. 2017. “Deep Keyphrase Generation.” arXiv. https://doi.org/10.48550/ARXIV.1704.06879.
Meng, Weibin, Ying Liu, Yichen Zhu, Shenglin Zhang, Dan Pei, Yuqing Liu, Yihao Chen, et al. 2019. “LogAnomaly: Unsupervised Detection of Sequential and Quantitative Anomalies in Unstructured Logs.” Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, August. https://doi.org/10.24963/ijcai.2019/658.
Meng, Wenjia, Zonghua Gu, Ming Zhang, and Zhaohui Wu. 2017. “Two-Bit Networks for Deep Learning on Resource-Constrained Embedded Devices.” arXiv. https://doi.org/10.48550/ARXIV.1701.00485.
Meng, Xiao-Li. 1994. “Multiple-Imputation Inferences with Uncongenial Sources of Input.” Statistical Science 9 (November). https://doi.org/10.1214/ss/1177010269.
Meng, Yen, Hsuan-Jui Chen, Jiatong Shi, Shinji Watanabe, Paola Garcia, Hung-yi Lee, and Hao Tang. 2022. “On Compressing Sequences for Self-Supervised Speech Models.” arXiv. https://doi.org/10.48550/ARXIV.2210.07189.
Meng, Zhong, Tongzhou Chen, Rohit Prabhavalkar, Yu Zhang, Gary Wang, Kartik Audhkhasi, Jesse Emond, et al. 2022. “Modular Hybrid Autoregressive Transducer.” arXiv. https://doi.org/10.48550/ARXIV.2210.17049.
Menghani, Gaurav. 2021. “Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better.” arXiv. https://doi.org/10.48550/ARXIV.2106.08962.
Merhav, Yuval, and Stephen Ash. 2018. “Design Challenges in Named Entity Transliteration.” arXiv. https://doi.org/10.48550/ARXIV.1808.02563.
Merity, Stephen, Nitish Shirish Keskar, and Richard Socher. 2017. “Regularizing and Optimizing LSTM Language Models.” arXiv. https://doi.org/10.48550/ARXIV.1708.02182.
Mesaros, Annamaria, Toni Heittola, Tuomas Virtanen, Emmanouil Benetos, Mathieu Lagrange, Grégoire Lafay, Peter Foster, and Mark D. Plumbley. 2017. “DCASE2016 Challenge Submissions Package,” September. https://doi.org/10.5281/ZENODO.926660.
Mescheder, Lars, Andreas Geiger, and Sebastian Nowozin. 2018. “Which Training Methods for GANs Do Actually Converge?” arXiv. https://doi.org/10.48550/ARXIV.1801.04406.
Mescheder, Lars, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. 2018. “Occupancy Networks: Learning 3D Reconstruction in Function Space.” arXiv. https://doi.org/10.48550/ARXIV.1812.03828.
Metz, Luke, Niru Maheswaranathan, Brian Cheung, and Jascha Sohl-Dickstein. 2018. “Meta-Learning Update Rules for Unsupervised Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1804.00222.
Metzen, Jan Hendrik, Tim Genewein, Volker Fischer, and Bastian Bischoff. 2017. “On Detecting Adversarial Perturbations.” arXiv. https://doi.org/10.48550/ARXIV.1702.04267.
Miao, Haoran, Gaofeng Cheng, Changfeng Gao, Pengyuan Zhang, and Yonghong Yan. 2020. “Transformer-Based Online CTC/Attention End-to-End Speech Recognition Architecture.” arXiv. https://doi.org/10.48550/ARXIV.2001.08290.
Miao, Hongyu, and Felix Xiaozhu Lin. 2021. “Enabling Large Neural Networks on Tiny Microcontrollers with Swapping.” arXiv. https://doi.org/10.48550/ARXIV.2101.08744.
Miao, Yajie, Mohammad Gowayyed, and Florian Metze. 2015. “EESEN: End-to-End Speech Recognition Using Deep RNN Models and WFST-Based Decoding.” arXiv. https://doi.org/10.48550/ARXIV.1507.08240.
Miao, Yukai, Yu Bai, Li Chen, Dan Li, Haifeng Sun, Xizheng Wang, Ziqiu Luo, et al. 2023. “An Empirical Study of NetOps Capability of Pre-Trained Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2309.05557.
Micenková, Barbora, Brian McWilliams, and Ira Assent. 2015. “Learning Representations for Outlier Detection on a Budget.” arXiv. https://doi.org/10.48550/ARXIV.1507.08104.
Miikkulainen, Risto, Jason Liang, Elliot Meyerson, Aditya Rawal, Dan Fink, Olivier Francon, Bala Raju, et al. 2017. “Evolving Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1703.00548.
Mikolov, Tomas, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. “Efficient Estimation of Word Representations in Vector Space.” arXiv. https://doi.org/10.48550/ARXIV.1301.3781.
Mikolov, Tomas, Edouard Grave, Piotr Bojanowski, Christian Puhrsch, and Armand Joulin. 2017. “Advances in Pre-Training Distributed Word Representations.” arXiv. https://doi.org/10.48550/ARXIV.1712.09405.
Mikolov, Tomas, Armand Joulin, and Marco Baroni. 2015. “A Roadmap Towards Machine Intelligence.” arXiv. https://doi.org/10.48550/ARXIV.1511.08130.
Mikolov, Tomas, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. “Distributed Representations of Words and Phrases and Their Compositionality.” arXiv. https://doi.org/10.48550/ARXIV.1310.4546.
Miller, Alexander, Adam Fisch, Jesse Dodge, Amir-Hossein Karimi, Antoine Bordes, and Jason Weston. 2016. “Key-Value Memory Networks for Directly Reading Documents.” arXiv. https://doi.org/10.48550/ARXIV.1606.03126.
Miller, Jason, and Scott Sheffield. 2016. “Imaginary Geometry i: Interacting SLEs.” Probability Theory and Related Fields 164 (March). https://doi.org/10.1007/s00440-016-0698-0.
———. 2017. “Imaginary Geometry IV: Interior Rays, Whole-Plane Reversibility, and Space-Filling Trees.” Probability Theory and Related Fields 169 (June). https://doi.org/10.1007/s00440-017-0780-2.
Milligan, Glenn W., and Martha C. Cooper. 1985. “An Examination of Procedures for Determining the Number of Clusters in a Data Set.” Psychometrika 50 (June). https://doi.org/10.1007/bf02294245.
Min, Bonan, Hayley Ross, Elior Sulem, Amir Pouran Ben Veyseh, Thien Huu Nguyen, Oscar Sainz, Eneko Agirre, Ilana Heinz, and Dan Roth. 2021. “Recent Advances in Natural Language Processing via Large Pre-Trained Language Models: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2111.01243.
Min, Seonwoo, Byunghan Lee, and Sungroh Yoon. 2016. “Deep Learning in Bioinformatics.” arXiv. https://doi.org/10.48550/ARXIV.1603.06430.
Minh, Anh Pham Thi, An Duc Nguyen, and Georgios Tzimiropoulos. 2023. “PRE: Vision-Language Prompt Learning with Reparameterization Encoder.” arXiv. https://doi.org/10.48550/ARXIV.2309.07760.
Miolane, Nina, Johan Mathe, Claire Donnat, Mikael Jorda, and Xavier Pennec. 2018. “Geomstats: A Python Package for Riemannian Geometry in Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1805.08308.
Mirza, Mehdi, and Simon Osindero. 2014. “Conditional Generative Adversarial Nets.” arXiv. https://doi.org/10.48550/ARXIV.1411.1784.
Mishchenko, Yuriy, Joshua T. Vogelstein, and Liam Paninski. 2011. “A Bayesian Approach for Inferring Neuronal Connectivity from Calcium Fluorescent Imaging Data.” The Annals of Applied Statistics 5 (June). https://doi.org/10.1214/09-aoas303.
Mishra, Abhinav, Ram Sriharsha, and Sichen Zhong. 2022. “OnlineSTL.” Proceedings of the VLDB Endowment 15 (March). https://doi.org/10.14778/3523210.3523219.
Mishra, Asit, Eriko Nurvitadhi, Jeffrey J Cook, and Debbie Marr. 2017. “WRPN: Wide Reduced-Precision Networks.” arXiv. https://doi.org/10.48550/ARXIV.1709.01134.
Mishra, Bhavana Dalvi, Oyvind Tafjord, and Peter Clark. 2022. “Towards Teachable Reasoning Systems: Using a Dynamic Memory of User Feedback for Continual System Improvement.” arXiv. https://doi.org/10.48550/ARXIV.2204.13074.
Misra, Ishan, and Laurens van der Maaten. 2019. “Self-Supervised Learning of Pretext-Invariant Representations.” arXiv. https://doi.org/10.48550/ARXIV.1912.01991.
Mitchell, Eric, Yoonho Lee, Alexander Khazatsky, Christopher D. Manning, and Chelsea Finn. 2023. “DetectGPT: Zero-Shot Machine-Generated Text Detection Using Probability Curvature.” arXiv. https://doi.org/10.48550/ARXIV.2301.11305.
Mitev, Richard, Anna Pazii, Markus Miettinen, William Enck, and Ahmad-Reza Sadeghi. 2020. “LeakyPick: IoT Audio Spy Detector.” Annual Computer Security Applications Conference, December. https://doi.org/10.1145/3427228.3427277.
Mitrović, Sandra, Davide Andreoletti, and Omran Ayoub. 2023. “ChatGPT or Human? Detect and Explain. Explaining Decisions of Machine Learning Model for Detecting Short ChatGPT-Generated Text.” arXiv. https://doi.org/10.48550/ARXIV.2301.13852.
Mittal, Anshul, Kunal Dahiya, Shreya Malani, Janani Ramaswamy, Seba Kuruvilla, Jitendra Ajmera, Keng-Hao Chang, Sumeet Agarwal, Purushottam Kar, and Manik Varma. 2022. “Multi-Modal Extreme Classification.” 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52688.2022.01207.
Mittal, Deepak, Shweta Bhardwaj, Mitesh M. Khapra, and Balaraman Ravindran. 2018. “Recovering from Random Pruning: On the Plasticity of Deep Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1801.10447.
Mitzenmacher, Michael, Rasmus Pagh, and Ninh Pham. 2014. “Efficient Estimation for High Similarities Using Odd Sketches.” Proceedings of the 23rd International Conference on World Wide Web, April. https://doi.org/10.1145/2566486.2568017.
Miyashita, Daisuke, Edward H. Lee, and Boris Murmann. 2016. “Convolutional Neural Networks Using Logarithmic Data Representation.” arXiv. https://doi.org/10.48550/ARXIV.1603.01025.
Miyato, Takeru, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. 2017. “Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning.” arXiv. https://doi.org/10.48550/ARXIV.1704.03976.
Mnih, Andriy, and Yee Whye Teh. 2012. “A Fast and Simple Algorithm for Training Neural Probabilistic Language Models.” arXiv. https://doi.org/10.48550/ARXIV.1206.6426.
Mnih, Volodymyr, Nicolas Heess, Alex Graves, and Koray Kavukcuoglu. 2014. “Recurrent Models of Visual Attention.” arXiv. https://doi.org/10.48550/ARXIV.1406.6247.
Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. 2013. “Playing Atari with Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1312.5602.
Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, et al. 2015. “Human-Level Control Through Deep Reinforcement Learning.” Nature 518 (February). https://doi.org/10.1038/nature14236.
Mo, Shentong, Enze Xie, Ruihang Chu, Lewei Yao, Lanqing Hong, Matthias Nießner, and Zhenguo Li. 2023. “DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation.” arXiv. https://doi.org/10.48550/ARXIV.2307.01831.
Mobaraki, Ehsan Bonabi, and Arijit Khan. 2023. “A Demonstration of Interpretability Methods for Graph Neural Networks.” Proceedings of the 6th Joint Workshop on Graph Data Management Experiences &Amp; Systems (GRADES) and Network Data Analytics (NDA), June. https://doi.org/10.1145/3594778.3594880.
“Modeling COVID-19 Scenarios for the United States.” 2020. Nature Medicine 27 (October). https://doi.org/10.1038/s41591-020-1132-9.
Modi, Ashutosh, Tatjana Anikina, Simon Ostermann, and Manfred Pinkal. 2017. “InScript: Narrative Texts Annotated with Script Information.” arXiv. https://doi.org/10.48550/ARXIV.1703.05260.
Mohamed, Shakir, and Balaji Lakshminarayanan. 2016. “Learning in Implicit Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.1610.03483.
Mohammadi, Mehdi, Ala Al-Fuqaha, Sameh Sorour, and Mohsen Guizani. 2017. “Deep Learning for IoT Big Data and Streaming Analytics: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.1712.04301.
Mohler, George. 2013. “Modeling and Estimation of Multi-Source Clustering in Crime and Security Data.” The Annals of Applied Statistics 7 (September). https://doi.org/10.1214/13-aoas647.
Moins, Théo, Julyan Arbel, Anne Dutfoy, and Stéphane Girard. 2023. “On the Use of a Local Rˆ to Improve MCMC Convergence Diagnostic.” Bayesian Analysis -1 (January). https://doi.org/10.1214/23-ba1399.
Mok, Tony C. W., and Albert C. S. Chung. 2019. “Learning Data Augmentation for Brain Tumor Segmentation with Coarse-to-Fine Generative Adversarial Networks.” Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries. https://doi.org/10.1007/978-3-030-11723-8_7.
Molchanov, Dmitry, Arsenii Ashukha, and Dmitry Vetrov. 2017. “Variational Dropout Sparsifies Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1701.05369.
Molchanov, Pavlo, Stephen Tyree, Tero Karras, Timo Aila, and Jan Kautz. 2016. “Pruning Convolutional Neural Networks for Resource Efficient Inference.” arXiv. https://doi.org/10.48550/ARXIV.1611.06440.
Mollá, Diego. 2023. “Large Language Models and Prompt Engineering for Biomedical Query Focused Multi-Document Summarisation.” arXiv. https://doi.org/10.48550/ARXIV.2311.05169.
Mondal, Shrayani, Rishabh Garodia, Arbaaz Qureshi, Taesung Lee, and Youngja Park. 2024. “Towards Generating Informative Textual Description for Neurons in Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.16731.
Monti, Federico, Davide Boscaini, Jonathan Masci, Emanuele Rodola, Jan Svoboda, and Michael M. Bronstein. 2017. “Geometric Deep Learning on Graphs and Manifolds Using Mixture Model CNNs.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July. https://doi.org/10.1109/cvpr.2017.576.
Monti, Federico, Davide Boscaini, Jonathan Masci, Emanuele Rodolà, Jan Svoboda, and Michael M. Bronstein. 2016. “Geometric Deep Learning on Graphs and Manifolds Using Mixture Model CNNs.” arXiv. https://doi.org/10.48550/ARXIV.1611.08402.
Moon, Kevin R., David van Dijk, Zheng Wang, Scott Gigante, Daniel B. Burkhardt, William S. Chen, Kristina Yim, et al. 2017. “Visualizing Structure and Transitions for Biological Data Exploration,” March. https://doi.org/10.1101/120378.
Moons, Philip, Koen Luyckx, Corina Thomet, Werner Budts, Junko Enomoto, Maayke A Sluman, Hsiao-Ling Yang, et al. 2022. “Patient-Reported Outcomes in the Aging Population of Adults with Congenital Heart Disease: Results from APPROACH-IS.” European Journal of Cardiovascular Nursing 22 (July). https://doi.org/10.1093/eurjcn/zvac057.
Moor, Michael, Oishi Banerjee, Zahra Shakeri Hossein Abad, Harlan M. Krumholz, Jure Leskovec, Eric J. Topol, and Pranav Rajpurkar. 2023. “Foundation Models for Generalist Medical Artificial Intelligence.” Nature 616 (April). https://doi.org/10.1038/s41586-023-05881-4.
Mordatch, Igor, and Pieter Abbeel. 2017. “Emergence of Grounded Compositional Language in Multi-Agent Populations.” arXiv. https://doi.org/10.48550/ARXIV.1703.04908.
Morgado, Antonio, Federico Heras, Mark Liffiton, Jordi Planes, and Joao Marques-Silva. 2013. “Iterative and Core-Guided MaxSAT Solving: A Survey and Assessment.” Constraints 18 (July). https://doi.org/10.1007/s10601-013-9146-2.
Morik, Marco, Ashudeep Singh, Jessica Hong, and Thorsten Joachims. 2020. “Controlling Fairness and Bias in Dynamic Learning-to-Rank.” Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3397271.3401100.
MORISE, Masanori, Fumiya YOKOMORI, and Kenji OZAWA. 2016. “WORLD: A Vocoder-Based High-Quality Speech Synthesis System for Real-Time Applications.” IEICE Transactions on Information and Systems E99.D. https://doi.org/10.1587/transinf.2015edp7457.
Moritz, Niko, Takaaki Hori, and Jonathan Le Roux. 2019. “Triggered Attention for End-to-End Speech Recognition.” ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), May. https://doi.org/10.1109/icassp.2019.8683510.
———. 2020. “Streaming Automatic Speech Recognition with the Transformer Model.” arXiv. https://doi.org/10.48550/ARXIV.2001.02674.
Mossel, Elchanan, Joe Neeman, and Allan Sly. 2014. “Reconstruction and Estimation in the Planted Partition Model.” Probability Theory and Related Fields 162 (July). https://doi.org/10.1007/s00440-014-0576-6.
Mou, Lili, Rui Men, Ge Li, Yan Xu, Lu Zhang, Rui Yan, and Zhi Jin. 2015. “Natural Language Inference by Tree-Based Convolution and Heuristic Matching.” arXiv. https://doi.org/10.48550/ARXIV.1512.08422.
Mousavi, S. Mostafa, Weiqiang Zhu, Yixiao Sheng, and Gregory C. Beroza. 2018. “CRED: A Deep Residual Network of Convolutional and Recurrent Units for Earthquake Signal Detection.” arXiv. https://doi.org/10.48550/ARXIV.1810.01965.
Mousouliotis, Panagiotis G., and Loukas P. Petrou. 2018. “SqueezeJet: High-Level Synthesis Accelerator Design for Deep Convolutional Neural Networks.” Applied Reconfigurable Computing. Architectures, Tools, and Applications. https://doi.org/10.1007/978-3-319-78890-6_5.
———. 2019. “Software-Defined FPGA Accelerator Design for Mobile Deep Learning Applications.” arXiv. https://doi.org/10.48550/ARXIV.1902.03192.
Moysset, Bastien, and Ronaldo Messina. 2019. “Are 2D-LSTM Really Dead for Offline Text Recognition?” International Journal on Document Analysis and Recognition (IJDAR) 22 (June). https://doi.org/10.1007/s10032-019-00325-0.
Mrazek, Vojtech, Zdenek Vasicek, Lukas Sekanina, Muhammad Abdullah Hanif, and Muhammad Shafique. 2019. “ALWANN: Automatic Layer-Wise Approximation of Deep Neural Network Accelerators Without Retraining.” 2019 IEEE/ACM International Conference on Computer-Aided Design (ICCAD), November. https://doi.org/10.1109/iccad45719.2019.8942068.
Mu, Jesse, Xiang Lisa Li, and Noah Goodman. 2023. “Learning to Compress Prompts with Gist Tokens.” arXiv. https://doi.org/10.48550/ARXIV.2304.08467.
Muandet, Krikamol, Arash Mehrjou, Si Kai Lee, and Anant Raj. 2019. “Dual Instrumental Variable Regression.” arXiv. https://doi.org/10.48550/ARXIV.1910.12358.
Muffo, Matteo, Aldo Cocco, and Enrico Bertino. 2023. “Evaluating Transformer Language Models on Arithmetic Operations Using Number Decomposition.” arXiv. https://doi.org/10.48550/ARXIV.2304.10977.
Muhammad, Shamsuddeen Hassan, Idris Abdulmumin, Abinew Ali Ayele, Nedjma Ousidhoum, David Ifeoluwa Adelani, Seid Muhie Yimam, Ibrahim Sa’id Ahmad, et al. 2023. “AfriSenti: A Twitter Sentiment Analysis Benchmark for African Languages.” arXiv. https://doi.org/10.48550/ARXIV.2302.08956.
Mukherjee, Indraneel, and Robert E. Schapire. 2011. “A Theory of Multiclass Boosting.” arXiv. https://doi.org/10.48550/ARXIV.1108.2989.
Müller, Robert, Fabian Ritz, Steffen Illium, and Claudia Linnhoff-Popien. 2021. “Acoustic Anomaly Detection for Machine Sounds Based on Image Transfer Learning.” Proceedings of the 13th International Conference on Agents and Artificial Intelligence. https://doi.org/10.5220/0010185800490056.
Müllner, Daniel. 2011. “Modern Hierarchical, Agglomerative Clustering Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.1109.2378.
Mummadi, Chaithanya Kumar, Tim Genewein, Dan Zhang, Thomas Brox, and Volker Fischer. 2019. “Group Pruning Using a Bounded-Lp Norm for Group Gating and Regularization.” arXiv. https://doi.org/10.48550/ARXIV.1908.03463.
Mun’im, Raden Mu’az, Nakamasa Inoue, and Koichi Shinoda. 2018. “Sequence-Level Knowledge Distillation for Model Compression of Attention-Based Sequence-to-Sequence Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1811.04531.
Muñoz-González, Luis, Battista Biggio, Ambra Demontis, Andrea Paudice, Vasin Wongrassamee, Emil C. Lupu, and Fabio Roli. 2017. “Towards Poisoning of Deep Learning Algorithms with Back-Gradient Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1708.08689.
Murdoch, W. James, and Arthur Szlam. 2017. “Automatic Rule Extraction from Long Short Term Memory Networks.” arXiv. https://doi.org/10.48550/ARXIV.1702.02540.
Murr, Lincoln, Morgan Grainger, and David Gao. 2023. “Testing LLMs on Code Generation with Varying Levels of Prompt Specificity.” arXiv. https://doi.org/10.48550/ARXIV.2311.07599.
Murray, Naila, and Albert Gordo. 2017. “A Deep Architecture for Unified Aesthetic Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1708.04890.
Mustafa, Ahmed, Jan Büthe, Srikanth Korse, Kishan Gupta, Guillaume Fuchs, and Nicola Pia. 2021. “A Streamwise GAN Vocoder for Wideband Speech Coding at Very Low Bit Rate.” arXiv. https://doi.org/10.48550/ARXIV.2108.04051.
Mustafa, Basil, Carlos Riquelme, Joan Puigcerver, Rodolphe Jenatton, and Neil Houlsby. 2022. “Multimodal Contrastive Learning with LIMoE: The Language-Image Mixture of Experts.” arXiv. https://doi.org/10.48550/ARXIV.2206.02770.
Myerscough, Keith, Jason Frank, and Benedict Leimkuhler. 2014. “Least-Biased Correction of Extended Dynamical Systems Using Observational Data.” arXiv. https://doi.org/10.48550/ARXIV.1411.6011.
Nachum, Ofir, Haoran Tang, Xingyu Lu, Shixiang Gu, Honglak Lee, and Sergey Levine. 2019. “Why Does Hierarchy (Sometimes) Work so Well in Reinforcement Learning?” arXiv. https://doi.org/10.48550/ARXIV.1909.10618.
Nagano, Yoshihiro, Shoichiro Yamaguchi, Yasuhiro Fujita, and Masanori Koyama. 2019. “A Wrapped Normal Distribution on Hyperbolic Space for Gradient-Based Learning.” arXiv. https://doi.org/10.48550/ARXIV.1902.02992.
Nakov, Preslav. 2021. “Improved Statistical Machine Translation Using Monolingual Paraphrases.” arXiv. https://doi.org/10.48550/ARXIV.2109.15119.
Nallapareddy, Vamsi, Nicola Bordin, Ian Sillitoe, Michael Heinzinger, Maria Littmann, Vaishali P Waman, Neeladri Sen, Burkhard Rost, and Christine Orengo. 2023. “CATHe: Detection of Remote Homologues for CATH Superfamilies Using Embeddings from Protein Language Models.” Bioinformatics 39 (January). https://doi.org/10.1093/bioinformatics/btad029.
Nallapati, Ramesh, Feifei Zhai, and Bowen Zhou. 2016. “SummaRuNNer: A Recurrent Neural Network Based Sequence Model for Extractive Summarization of Documents.” arXiv. https://doi.org/10.48550/ARXIV.1611.04230.
Nandanwar, Sharad, and M. N. Murty. 2016. “Structural Neighborhood Based Classification of Nodes in a Network.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939782.
Narasimhan, Karthik, Tejas Kulkarni, and Regina Barzilay. 2015. “Language Understanding for Text-Based Games Using Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1506.08941.
Narayanan, Annamalai, Mahinthan Chandramohan, Lihui Chen, Yang Liu, and Santhoshkumar Saminathan. 2016. “Subgraph2vec: Learning Distributed Representations of Rooted Sub-Graphs from Large Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1606.08928.
Navigli, R., and M. Lapata. 2010. “An Experimental Study of Graph Connectivity for Unsupervised Word Sense Disambiguation.” IEEE Transactions on Pattern Analysis and Machine Intelligence 32 (April). https://doi.org/10.1109/tpami.2009.36.
Nazary, Fatemeh, Yashar Deldjoo, and Tommaso Di Noia. 2023. “ChatGPT-HealthPrompt. Harnessing the Power of XAI in Prompt-Based Healthcare Decision Support Using ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2308.09731.
Neal, Radford M. 2003. “Slice Sampling.” The Annals of Statistics 31 (June). https://doi.org/10.1214/aos/1056562461.
Nedelkoski, Sasho, Jasmin Bogatinovski, Ajay Kumar Mandapati, and Jorge Cardoso. 2019. “Multi-Source Distributed System Data for AI-Powered Analytics,” October. https://doi.org/10.5281/ZENODO.3484800.
Neelakantan, Arvind, Benjamin Roth, and Andrew McCallum. 2015. “Compositional Vector Space Models for Knowledge Base Completion.” arXiv. https://doi.org/10.48550/ARXIV.1504.06662.
Neelakantan, Arvind, Jeevan Shankar, Alexandre Passos, and Andrew McCallum. 2015. “Efficient Non-Parametric Estimation of Multiple Embeddings Per Word in Vector Space.” arXiv. https://doi.org/10.48550/ARXIV.1504.06654.
Negahbani, Farzin, Rasool Sabzi, Bita Pakniyat Jahromi, Dena Firouzabadi, Fateme Movahedi, Mahsa Kohandel Shirazi, Shayan Majidi, and Amirreza Dehghanian. 2021. “PathoNet Introduced as a Deep Neural Network Backend for Evaluation of Ki-67 and Tumor-Infiltrating Lymphocytes in Breast Cancer.” Scientific Reports 11 (April). https://doi.org/10.1038/s41598-021-86912-w.
NEIL, MARTIN, NORMAN FENTON, and LARS NIELSON. 2000. “Building Large-Scale Bayesian Networks.” The Knowledge Engineering Review 15 (September). https://doi.org/10.1017/s0269888900003039.
Nelson, John. 2020. “Multivariate Mapping.” Geographic Information Science &Amp; Technology Body of Knowledge 2020 (January). https://doi.org/10.22224/gistbok/2020.1.5.
Nelson, Jonathan D. 2005. “Finding Useful Questions: On Bayesian Diagnosticity, Probability, Impact, and Information Gain.” Psychological Review 112. https://doi.org/10.1037/0033-295x.112.4.979.
Neubig, Graham. 2017. “Neural Machine Translation and Sequence-to-Sequence Models: A Tutorial.” arXiv. https://doi.org/10.48550/ARXIV.1703.01619.
Neubig, Graham, Chris Dyer, Yoav Goldberg, Austin Matthews, Waleed Ammar, Antonios Anastasopoulos, Miguel Ballesteros, et al. 2017. “DyNet: The Dynamic Neural Network Toolkit.” arXiv. https://doi.org/10.48550/ARXIV.1701.03980.
Neubig, Graham, and Junjie Hu. 2018. “Rapid Adaptation of Neural Machine Translation to New Languages.” arXiv. https://doi.org/10.48550/ARXIV.1808.04189.
Neumann, Luka, and Jiri Matas. 2013. “Scene Text Localization and Recognition with Oriented Stroke Detection.” 2013 IEEE International Conference on Computer Vision, December. https://doi.org/10.1109/iccv.2013.19.
Neumann, Michael, and Ngoc Thang Vu. 2017. “Attentive Convolutional Neural Network Based Speech Emotion Recognition: A Study on the Impact of Input Features, Signal Length, and Acted Speech.” arXiv. https://doi.org/10.48550/ARXIV.1706.00612.
Ng, Joe Yue-Hei, Matthew Hausknecht, Sudheendra Vijayanarasimhan, Oriol Vinyals, Rajat Monga, and George Toderici. 2015. “Beyond Short Snippets: Deep Networks for Video Classification.” arXiv. https://doi.org/10.48550/ARXIV.1503.08909.
Ngo, Tuan Duc, Binh-Son Hua, and Khoi Nguyen. 2023. “ISBNet: A 3D Point Cloud Instance Segmentation Network with Instance-Aware Sampling and Box-Aware Dynamic Convolution.” arXiv. https://doi.org/10.48550/ARXIV.2303.00246.
Nguyen, Dat Quoc. 2017. “A Survey of Embedding Models of Entities and Relationships for Knowledge Graph Completion.” arXiv. https://doi.org/10.48550/ARXIV.1703.08098.
Nguyen, Thai-Son, Ngoc-Quan Pham, Sebastian Stueker, and Alex Waibel. 2020. “High Performance Sequence-to-Sequence Model for Streaming Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2003.10022.
Nguyen, Van-Hoang, Kazunari Sugiyama, Preslav Nakov, and Min-Yen Kan. 2020. “FANG.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412046.
Ni, Haomiao, Changhao Shi, Kai Li, Sharon X. Huang, and Martin Renqiang Min. 2023. “Conditional Image-to-Video Generation with Latent Flow Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.13744.
Nichol, Alex, Joshua Achiam, and John Schulman. 2018. “On First-Order Meta-Learning Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.1803.02999.
Nickel, Maximilian, Volker Tresp, and Hans-Peter Kriegel. 2012. “Factorizing YAGO.” Proceedings of the 21st International Conference on World Wide Web, April. https://doi.org/10.1145/2187836.2187874.
Nicolau, Marcio, Anderson R. Tavares, Zhiwei Zhang, Pedro Avelar, João M. Flach, Luis C. Lamb, and Moshe Y. Vardi. 2020. “Understanding Boolean Function Learnability on Deep Neural Networks: PAC Learning Meets Neurosymbolic Models.” arXiv. https://doi.org/10.48550/ARXIV.2009.05908.
Nigam, Priyanka, Yiwei Song, Vijai Mohan, Vihan Lakshman, Weitian (Allen) Ding, Ankit Shingavi, Choon Hui Teo, Hao Gu, and Bing Yin. 2019. “Semantic Product Search.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330759.
Nijkamp, Erik, Jeffrey Ruffolo, Eli N. Weinstein, Nikhil Naik, and Ali Madani. 2022. “ProGen2: Exploring the Boundaries of Protein Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2206.13517.
Niknam, Solmaz, Harpreet S. Dhillon, and Jeffery H. Reed. 2019. “Federated Learning for Wireless Communications: Motivation, Opportunities and Challenges.” arXiv. https://doi.org/10.48550/ARXIV.1908.06847.
Nikolentzos, Giannis, Polykarpos Meladianos, Stratis Limnios, and Michalis Vazirgiannis. 2018. “A Degeneracy Framework for Graph Similarity.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/360.
Nishimura, Masanari, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda. 2016. “Singing Voice Synthesis Based on Deep Neural Networks.” Interspeech 2016, September. https://doi.org/10.21437/interspeech.2016-1027.
Niu, Changan, Chuanyi Li, Vincent Ng, Dongxiao Chen, Jidong Ge, and Bin Luo. 2023. “An Empirical Comparison of Pre-Trained Models of Source Code.” arXiv. https://doi.org/10.48550/ARXIV.2302.04026.
Niu, Feng, Benjamin Recht, Christopher Re, and Stephen J. Wright. 2011. “HOGWILD!: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent.” arXiv. https://doi.org/10.48550/ARXIV.1106.5730.
Nobata, Chikashi, Joel Tetreault, Achint Thomas, Yashar Mehdad, and Yi Chang. 2016. “Abusive Language Detection in Online User Content.” Proceedings of the 25th International Conference on World Wide Web, April. https://doi.org/10.1145/2872427.2883062.
Nodelman, Uri, Christian R. Shelton, and Daphne Koller. 2013. “Continuous Time Bayesian Networks.” arXiv. https://doi.org/10.48550/ARXIV.1301.0591.
Noè, Umberto, Alan Lazarus, Hao Gao, Vinny Davies, Benn Macdonald, Kenneth Mangion, Colin Berry, Xiaoyu Luo, and Dirk Husmeier. 2019. “Gaussian Process Emulation to Accelerate Parameter Estimation in a Mechanical Model of the Left Ventricle: A Critical Step Towards Clinical End-User Relevance.” Journal of The Royal Society Interface 16 (July). https://doi.org/10.1098/rsif.2019.0114.
Nori, Harsha, Nicholas King, Scott Mayer McKinney, Dean Carignan, and Eric Horvitz. 2023. “Capabilities of GPT-4 on Medical Challenge Problems.” arXiv. https://doi.org/10.48550/ARXIV.2303.13375.
Noroozi, Mehdi, and Paolo Favaro. 2016. “Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles.” arXiv. https://doi.org/10.48550/ARXIV.1603.09246.
North, Joshua S., Christopher K. Wikle, and Erin M. Schliep. 2022. “A Bayesian Approach for Spatio-Temporal Data-Driven Dynamic Equation Discovery.” arXiv. https://doi.org/10.48550/ARXIV.2209.02750.
Notaro, Paolo, Jorge Cardoso, and Michael Gerndt. 2021. “A Survey of AIOps Methods for Failure Management.” ACM Transactions on Intelligent Systems and Technology 12 (November). https://doi.org/10.1145/3483424.
Nottingham, Kolby, Prithviraj Ammanabrolu, Alane Suhr, Yejin Choi, Hannaneh Hajishirzi, Sameer Singh, and Roy Fox. 2023. “Do Embodied Agents Dream of Pixelated Sheep: Embodied Decision Making Using Language Guided World Modelling.” arXiv. https://doi.org/10.48550/ARXIV.2301.12050.
Nov, Oded, Nina Singh, and Devin M. Mann. 2023. “Putting ChatGPT’s Medical Advice to the (Turing) Test,” January. https://doi.org/10.1101/2023.01.23.23284735.
Nowak, Theodore S., and Jason J. Corso. 2018. “Deep Net Triage: Analyzing the Importance of Network Layers via Structural Compression.” arXiv. https://doi.org/10.48550/ARXIV.1801.04651.
Nowak-Vila, Alex, Francis Bach, and Alessandro Rudi. 2020. “Consistent Structured Prediction with Max-Min Margin Markov Networks.” arXiv. https://doi.org/10.48550/ARXIV.2007.01012.
Noy, Asaf, Niv Nayman, Tal Ridnik, Nadav Zamir, Sivan Doveh, Itamar Friedman, Raja Giryes, and Lihi Zelnik-Manor. 2019. “ASAP: Architecture Search, Anneal and Prune.” arXiv. https://doi.org/10.48550/ARXIV.1904.04123.
Nualart, D., and E. Pardoux. 1988. “Stochastic Calculus with Anticipating Integrands.” Probability Theory and Related Fields 78 (August). https://doi.org/10.1007/bf00353876.
Numeroso, Danilo, and Davide Bacciu. 2020. “Explaining Deep Graph Networks with Molecular Counterfactuals.” arXiv. https://doi.org/10.48550/ARXIV.2011.05134.
O’Leary, Jared, Joel A. Paulson, and Ali Mesbah. 2022. “Stochastic Physics-Informed Neural Ordinary Differential Equations.” Journal of Computational Physics 468 (November). https://doi.org/10.1016/j.jcp.2022.111466.
O’Shea, Keiron, and Ryan Nash. 2015. “An Introduction to Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.08458.
O’Shea, Timothy J, Kiran Karra, and T. Charles Clancy. 2016. “Learning to Communicate: Channel Auto-Encoders, Domain Specific Regularizers, and Attention.” arXiv. https://doi.org/10.48550/ARXIV.1608.06409.
Obukhov, Anton, Elvis Yu-Jing Lin, Jonathan Kyl, and Semen Zhydenko. 2021. “Toshas/Torch-Fidelity: Version 0.3.0,” June. https://doi.org/10.5281/ZENODO.4957738.
Ochiai, Tsubasa, Shinji Watanabe, Takaaki Hori, and John R. Hershey. 2017. “Multichannel End-to-End Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1703.04783.
Odena, Augustus, Christopher Olah, and Jonathon Shlens. 2016. “Conditional Image Synthesis with Auxiliary Classifier GANs.” arXiv. https://doi.org/10.48550/ARXIV.1610.09585.
Oguntola, Ini, Subby Olubeko, and Christopher Sweeney. 2018. “SlimNets: An Exploration of Deep Model Compression and Acceleration.” arXiv. https://doi.org/10.48550/ARXIV.1808.00496.
Oh, Hyun-Jung, Ana Muriel, Hari Balasubramanian, Katherine Atkinson, and Thomas Ptaszkiewicz. 2013. “Guidelines for Scheduling in Primary Care Under Different Patient Types and Stochastic Nurse and Provider Service Times.” IIE Transactions on Healthcare Systems Engineering 3 (October). https://doi.org/10.1080/19488300.2013.858379.
Oliehoek, Frans A., and Christopher Amato. 2016. “A Concise Introduction to Decentralized POMDPs.” SpringerBriefs in Intelligent Systems. https://doi.org/10.1007/978-3-319-28929-8.
Oliveira, Catarina, João Torres, Maria Inês Silva, David Aparício, João Tiago Ascensão, and Pedro Bizarro. 2021. “GuiltyWalker: Distance to Illicit Nodes in the Bitcoin Network.” arXiv. https://doi.org/10.48550/ARXIV.2102.05373.
Ollivier, Yann, Ludovic Arnold, Anne Auger, and Nikolaus Hansen. 2011. “Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles.” arXiv. https://doi.org/10.48550/ARXIV.1106.3708.
Omar, Reham, Omij Mangukiya, Panos Kalnis, and Essam Mansour. 2023. “ChatGPT Versus Traditional Question Answering for Knowledge Graphs: Current Status and Future Directions Towards Knowledge Graph Chatbots.” arXiv. https://doi.org/10.48550/ARXIV.2302.06466.
Omidshafiei, Shayegan, Jason Pazis, Christopher Amato, Jonathan P. How, and John Vian. 2017. “Deep Decentralized Multi-Task Multi-Agent Reinforcement Learning Under Partial Observability.” arXiv. https://doi.org/10.48550/ARXIV.1703.06182.
Oñoro-Rubio, Daniel, Mathias Niepert, Alberto García-Durán, Roberto González, and Roberto J. López-Sastre. 2017. “Answering Visual-Relational Queries in Web-Extracted Knowledge Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1709.02314.
Oord, Aaron van den, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner, Andrew Senior, and Koray Kavukcuoglu. 2016. “WaveNet: A Generative Model for Raw Audio.” arXiv. https://doi.org/10.48550/ARXIV.1609.03499.
Oord, Aaron van den, Nal Kalchbrenner, and Koray Kavukcuoglu. 2016. “Pixel Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1601.06759.
Oord, Aaron van den, Nal Kalchbrenner, Oriol Vinyals, Lasse Espeholt, Alex Graves, and Koray Kavukcuoglu. 2016. “Conditional Image Generation with PixelCNN Decoders.” arXiv. https://doi.org/10.48550/ARXIV.1606.05328.
Oord, Aaron van den, Yazhe Li, Igor Babuschkin, Karen Simonyan, Oriol Vinyals, Koray Kavukcuoglu, George van den Driessche, et al. 2017. “Parallel WaveNet: Fast High-Fidelity Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1711.10433.
Oord, Aaron van den, Oriol Vinyals, and Koray Kavukcuoglu. 2017. “Neural Discrete Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1711.00937.
Oosterhuis, Harrie. 2021. “Computationally Efficient Optimization of Plackett-Luce Ranking Models for Relevance and Fairness.” Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3404835.3462830.
“Open Source Indicators Project.” 2017. https://doi.org/10.7910/DVN/EN8FUW.
Oprea, Silviu, and Walid Magdy. 2019. “Exploring Author Context for Detecting Intended Vs Perceived Sarcasm.” Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/p19-1275.
Orbes-Arteaga, Mauricio, Jorge Cardoso, Lauge Sørensen, Christian Igel, Sebastien Ourselin, Marc Modat, Mads Nielsen, and Akshay Pai. 2019. “Knowledge Distillation for Semi-Supervised Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1908.07355.
Ore, John-Paul, Sebastian Elbaum, Carrick Detweiler, and Lambros Karkazis. 2018. “Assessing the Type Annotation Burden.” Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering, September. https://doi.org/10.1145/3238147.3238173.
Oreshkin, Boris N., Dmitri Carpov, Nicolas Chapados, and Yoshua Bengio. 2020. “Meta-Learning Framework with Applications to Zero-Shot Time-Series Forecasting.” arXiv. https://doi.org/10.48550/ARXIV.2002.02887.
Orihashi, Shota, Yoshihiro Yamazaki, Naoki Makishima, Mana Ihori, Akihiko Takashima, Tomohiro Tanaka, and Ryo Masumura. 2021. “Utilizing Resource-Rich Language Datasets for End-to-End Scene Text Recognition in Resource-Poor Languages.” ACM Multimedia Asia, December. https://doi.org/10.1145/3469877.3490571.
“Orion-Wyc/GAGA: Release for the WWW23 Artifacts Available.” 2023, February. https://doi.org/10.5281/ZENODO.7608986.
Orsingher, Enzo, and Luisa Beghin. 2003. “Time-Fractional Telegraph Equations and Telegraph Processes with Brownian Time.” Probability Theory and Related Fields 128 (November). https://doi.org/10.1007/s00440-003-0309-8.
Ortega, Antonio, Pascal Frossard, Jelena Kovačević, José M. F. Moura, and Pierre Vandergheynst. 2017. “Graph Signal Processing: Overview, Challenges and Applications.” arXiv. https://doi.org/10.48550/ARXIV.1712.00468.
Ortiz-Jimenez, Guillermo, Apostolos Modas, Seyed-Mohsen Moosavi-Dezfooli, and Pascal Frossard. 2020. “Neural Anisotropy Directions.” arXiv. https://doi.org/10.48550/ARXIV.2006.09717.
Oseledets, I. V. 2011. “Tensor-Train Decomposition.” SIAM Journal on Scientific Computing 33 (January). https://doi.org/10.1137/090752286.
Ostendorff, Malte, Peter Bourgonje, Maria Berger, Julian Moreno-Schneider, Georg Rehm, and Bela Gipp. 2019. “Enriching BERT with Knowledge Graph Embeddings for Document Classification.” arXiv. https://doi.org/10.48550/ARXIV.1909.08402.
Oswald, Johannes von, Eyvind Niklasson, Ettore Randazzo, João Sacramento, Alexander Mordvintsev, Andrey Zhmoginov, and Max Vladymyrov. 2022. “Transformers Learn in-Context by Gradient Descent.” arXiv. https://doi.org/10.48550/ARXIV.2212.07677.
Otter, Daniel W., Julian R. Medina, and Jugal K. Kalita. 2018. “A Survey of the Usages of Deep Learning in Natural Language Processing.” arXiv. https://doi.org/10.48550/ARXIV.1807.10854.
Ou, Weitong, Bo Chen, Xinyi Dai, Weinan Zhang, Weiwen Liu, Ruiming Tang, and Yong Yu. 2023. “A Survey on Bid Optimization in Real-Time Bidding Display Advertising.” ACM Transactions on Knowledge Discovery from Data 18 (December). https://doi.org/10.1145/3628603.
Ouyang, Wentao, Xiuwu Zhang, Lei Zhao, Jinmei Luo, Yu Zhang, Heng Zou, Zhaojie Liu, and Yanlong Du. 2020. “MiNet.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412728.
“Overview of Industrial Batch Process Scheduling.” 2010. Chemical Engineering Transactions 21 (September). https://doi.org/10.3303/CET1021150.
Owhadi, Houman. 2015. “Bayesian Numerical Homogenization.” Multiscale Modeling &Amp; Simulation 13 (January). https://doi.org/10.1137/140974596.
Oymak, Samet, Amin Jalali, Maryam Fazel, Yonina C. Eldar, and Babak Hassibi. 2012. “Simultaneously Structured Models with Application to Sparse and Low-Rank Matrices.” arXiv. https://doi.org/10.48550/ARXIV.1212.3753.
Ozcaglar, Cagri, Sahin Geyik, Brian Schmitz, Prakhar Sharma, Alex Shelkovnykov, Yiming Ma, and Erik Buchanan. 2019. “Entity Personalized Talent Search Models with Tree Interaction Features.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313672.
Paganini, Michela. 2017. “Pythia Generated Jet Images for Location Aware Generative Adversarial Network Training,” February. https://doi.org/10.17632/4R4V785RGX.1.
Painsky, Amichai, and Saharon Rosset. 2018. “Lossless (and Lossy) Compression of Random Forests.” arXiv. https://doi.org/10.48550/ARXIV.1810.11197.
Pal, Aditya, Chantat Eksombatchai, Yitong Zhou, Bo Zhao, Charles Rosenberg, and Jure Leskovec. 2020. “PinnerSage.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403280.
Paleyes, Andrei, Raoul-Gabriel Urma, and Neil D. Lawrence. 2022. “Challenges in Deploying Machine Learning: A Survey of Case Studies.” ACM Computing Surveys 55 (December). https://doi.org/10.1145/3533378.
Palmero, Cristina, Javier Selva, Sorina Smeureanu, Julio C. S. Jacques Junior, Albert Clapés, Alexa Moseguí, Zejian Zhang, et al. 2020. “Context-Aware Personality Inference in Dyadic Scenarios: Introducing the UDIVA Dataset.” arXiv. https://doi.org/10.48550/ARXIV.2012.14259.
Pan, Boyuan, Hao Li, Zhou Zhao, Bin Cao, Deng Cai, and Xiaofei He. 2017. “MEMEN: Multi-Layer Embedding with Memory Networks for Machine Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.1707.09098.
Pan, Junting, Cristian Canton Ferrer, Kevin McGuinness, Noel E. O’Connor, Jordi Torres, Elisa Sayrol, and Xavier Giro-i-Nieto. 2017. “SalGAN: Visual Saliency Prediction with Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1701.01081.
Pan, Junwei, Jian Xu, Alfonso Lobos Ruiz, Wenliang Zhao, Shengjun Pan, Yu Sun, and Quan Lu. 2018. “Field-Weighted Factorization Machines for Click-Through Rate Prediction in Display Advertising.” Proceedings of the 2018 World Wide Web Conference on World Wide Web - WWW ’18. https://doi.org/10.1145/3178876.3186040.
Pan, Rui, Shuo Xing, Shizhe Diao, Xiang Liu, Kashun Shum, Jipeng Zhang, and Tong Zhang. 2023. “Plum: Prompt Learning Using Metaheuristic.” arXiv. https://doi.org/10.48550/ARXIV.2311.08364.
Pan, Shirui, Ruiqi Hu, Guodong Long, Jing Jiang, Lina Yao, and Chengqi Zhang. 2018. “Adversarially Regularized Graph Autoencoder for Graph Embedding.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/362.
Pan, Shirui, Linhao Luo, Yufei Wang, Chen Chen, Jiapu Wang, and Xindong Wu. 2024. “Unifying Large Language Models and Knowledge Graphs: A Roadmap.” IEEE Transactions on Knowledge and Data Engineering. https://doi.org/10.1109/tkde.2024.3352100.
Pan, Xinghao, Shivaram Venkataraman, Zizheng Tai, and Joseph Gonzalez. 2017. “Hemingway: Modeling Distributed Optimization Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.1702.05865.
Panchendrarajan, Rrubaa, and Arkaitz Zubiaga. 2024. “Synergizing Machine Learning &Amp; Symbolic Methods: A Survey on Hybrid Approaches to Natural Language Processing.” arXiv. https://doi.org/10.48550/ARXIV.2401.11972.
Pang, Bo, Lillian Lee, and Shivakumar Vaithyanathan. 2002. “Thumbs Up?” Proceedings of the ACL-02 Conference on Empirical Methods in Natural Language Processing - EMNLP ’02. https://doi.org/10.3115/1118693.1118704.
Pang, Guansong, Longbing Cao, Ling Chen, and Huan Liu. 2018. “Learning Representations of Ultrahigh-Dimensional Data for Random Distance-Based Outlier Detection.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220042.
Pang, Tianyu, Kun Xu, Chao Du, Ning Chen, and Jun Zhu. 2019. “Improving Adversarial Robustness via Promoting Ensemble Diversity.” arXiv. https://doi.org/10.48550/ARXIV.1901.08846.
Paninski, Liam. 2003. “Estimation of Entropy and Mutual Information.” Neural Computation 15 (June). https://doi.org/10.1162/089976603321780272.
Papageorgiou, Ioannis, and Ioannis Kontoyiannis. 2023. “Posterior Representations for Bayesian Context Trees: Sampling, Estimation and Convergence.” Bayesian Analysis -1 (January). https://doi.org/10.1214/23-ba1362.
Paparrizos, John, and Luis Gravano. 2015. “K-Shape.” Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, May. https://doi.org/10.1145/2723372.2737793.
Paparrizos, John, Yuhao Kang, Paul Boniol, Ruey S. Tsay, Themis Palpanas, and Michael J. Franklin. 2022. “TSB-UAD.” Proceedings of the VLDB Endowment 15 (April). https://doi.org/10.14778/3529337.3529354.
Papernot, Nicolas, Martín Abadi, Úlfar Erlingsson, Ian Goodfellow, and Kunal Talwar. 2016. “Semi-Supervised Knowledge Transfer for Deep Learning from Private Training Data.” arXiv. https://doi.org/10.48550/ARXIV.1610.05755.
Papernot, Nicolas, Fartash Faghri, Nicholas Carlini, Ian Goodfellow, Reuben Feinman, Alexey Kurakin, Cihang Xie, et al. 2016. “Technical Report on the CleverHans V2.1.0 Adversarial Examples Library.” arXiv. https://doi.org/10.48550/ARXIV.1610.00768.
Papernot, Nicolas, Patrick McDaniel, Ananthram Swami, and Richard Harang. 2016. “Crafting Adversarial Input Sequences for Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1604.08275.
Papernot, Nicolas, Patrick McDaniel, Xi Wu, Somesh Jha, and Ananthram Swami. 2015. “Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1511.04508.
Parapar, Javier, and Filip Radlinski. 2021. “Towards Unified Metrics for Accuracy and Diversity for Recommender Systems.” Fifteenth ACM Conference on Recommender Systems, September. https://doi.org/10.1145/3460231.3474234.
Parashar, Angshuman, Minsoo Rhu, Anurag Mukkara, Antonio Puglielli, Rangharajan Venkatesan, Brucek Khailany, Joel Emer, Stephen W. Keckler, and William J. Dally. 2017. “SCNN: An Accelerator for Compressed-Sparse Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1708.04485.
Pardoux, Etienne, and Shige Peng. 1994. “Backward Doubly Stochastic Differential Equations and Systems of Quasilinear SPDEs.” Probability Theory and Related Fields 98 (June). https://doi.org/10.1007/bf01192514.
Pardoux, Etienne, and Shanjian Tang. 1999. “Forward-Backward Stochastic Differential Equations and Quasilinear Parabolic PDEs.” Probability Theory and Related Fields 114 (May). https://doi.org/10.1007/s004409970001.
Parisotto, Emilio, Jimmy Lei Ba, and Ruslan Salakhutdinov. 2015. “Actor-Mimic: Deep Multitask and Transfer Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1511.06342.
Parisotto, Emilio, Abdel-rahman Mohamed, Rishabh Singh, Lihong Li, Dengyong Zhou, and Pushmeet Kohli. 2016. “Neuro-Symbolic Program Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1611.01855.
Park, Daehyung, Yuuna Hoshi, and Charles C. Kemp. 2017. “A Multimodal Anomaly Detector for Robot-Assisted Feeding Using an LSTM-Based Variational Autoencoder.” arXiv. https://doi.org/10.48550/ARXIV.1711.00614.
Park, Daniel S., William Chan, Yu Zhang, Chung-Cheng Chiu, Barret Zoph, Ekin D. Cubuk, and Quoc V. Le. 2019. “SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition.” Interspeech 2019, September. https://doi.org/10.21437/interspeech.2019-2680.
Park, Hogun, and Jennifer Neville. 2019. “Exploiting Interaction Links for Node Classification with Deep Graph Neural Networks.” Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, August. https://doi.org/10.24963/ijcai.2019/447.
Park, Jihyun, Kexin Zhao, Kainan Peng, and Wei Ping. 2019. “Multi-Speaker End-to-End Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1907.04462.
Park, Joon Sung, Joseph O’Brien, Carrie Jun Cai, Meredith Ringel Morris, Percy Liang, and Michael S. Bernstein. 2023. “Generative Agents: Interactive Simulacra of Human Behavior.” Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, October. https://doi.org/10.1145/3586183.3606763.
Park, Namyong, Andrey Kan, Xin Luna Dong, Tong Zhao, and Christos Faloutsos. 2019. “Estimating Node Importance in Knowledge Graphs Using Graph Neural Networks.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330855.
Park, SeongUk, and Nojun Kwak. 2019. “FEED: Feature-Level Ensemble for Knowledge Distillation.” arXiv. https://doi.org/10.48550/ARXIV.1909.10754.
Park, Sung Min, Kristian Georgiev, Andrew Ilyas, Guillaume Leclerc, and Aleksander Madry. 2023. “TRAK: Attributing Model Behavior at Scale.” arXiv. https://doi.org/10.48550/ARXIV.2303.14186.
Park, Yong-chan, Jun-Gi Jang, and U Kang. 2021. “Fast and Accurate Partial Fourier Transform for Time Series Data.” Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3447548.3467293.
Parrilo, Pablo A. 2003. “Semidefinite Programming Relaxations for Semialgebraic Problems.” Mathematical Programming 96 (May). https://doi.org/10.1007/s10107-003-0387-5.
Pasini, Tommaso, Francesco Maria Elia, and Roberto Navigli. 2018. “Huge Automatically Extracted Training Sets for Multilingual Word Sense Disambiguation.” arXiv. https://doi.org/10.48550/ARXIV.1805.04685.
Paszke, Adam, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, et al. 2019. “PyTorch: An Imperative Style, High-Performance Deep Learning Library.” arXiv. https://doi.org/10.48550/ARXIV.1912.01703.
Patel, Komal. 2017. “Testing the Limits of the First Amendment: How a CFAA Prohibition on Online Antidiscrimination Testing Infringes on Protected Speech Activity.” SSRN Electronic Journal. https://doi.org/10.2139/ssrn.3046847.
Patel, Sajan B, and Kyle Lam. 2023. “ChatGPT: The Future of Discharge Summaries?” The Lancet Digital Health 5 (March). https://doi.org/10.1016/s2589-7500(23)00021-3.
Pathak, Deepak, Ross Girshick, Piotr Dollár, Trevor Darrell, and Bharath Hariharan. 2016. “Learning Features by Watching Objects Move.” arXiv. https://doi.org/10.48550/ARXIV.1612.06370.
“Pattern Recognition.” 2017. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-319-66709-6.
Paul, Debjit, Mete Ismayilzada, Maxime Peyrard, Beatriz Borges, Antoine Bosselut, Robert West, and Boi Faltings. 2023. “REFINER: Reasoning Feedback on Intermediate Representations.” arXiv. https://doi.org/10.48550/ARXIV.2304.01904.
Paul, Rishov, Md. Mohib Hossain, Mohammed Latif Siddiq, Masum Hasan, Anindya Iqbal, and Joanna C. S. Santos. 2023. “Enhancing Automated Program Repair Through Fine-Tuning and Prompt Engineering.” arXiv. https://doi.org/10.48550/ARXIV.2304.07840.
Paulus, Romain, Caiming Xiong, and Richard Socher. 2017. “A Deep Reinforced Model for Abstractive Summarization.” arXiv. https://doi.org/10.48550/ARXIV.1705.04304.
Pazho, Armin Danesh, Ghazal Alinezhad Noghre, Arnab A Purkayastha, Jagannadh Vempati, Otto Martin, and Hamed Tabkhi. 2022. “A Survey of Graph-Based Deep Learning for Anomaly Detection in Distributed Systems.” arXiv. https://doi.org/10.48550/ARXIV.2206.04149.
Pearce, Hammond, Benjamin Tan, Baleegh Ahmad, Ramesh Karri, and Brendan Dolan-Gavitt. 2021. “Examining Zero-Shot Vulnerability Repair with Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2112.02125.
PEARSON, KARL. 1905. “The Problem of the Random Walk.” Nature 72 (July). https://doi.org/10.1038/072294b0.
Péché, S. 2005. “The Largest Eigenvalue of Small Rank Perturbations of Hermitian Random Matrices.” Probability Theory and Related Fields 134 (August). https://doi.org/10.1007/s00440-005-0466-z.
Pedregosa, Fabian, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, et al. 2012. “Scikit-Learn: Machine Learning in Python.” arXiv. https://doi.org/10.48550/ARXIV.1201.0490.
Peled, Lotem, and Roi Reichart. 2017. “Sarcasm SIGN: Interpreting Sarcasm with Sentiment Based Monolingual Machine Translation.” Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/p17-1155.
Peña, Fidel A. Guerrero, Heitor Rapela Medeiros, Thomas Dubail, Masih Aminbeidokhti, Eric Granger, and Marco Pedersoli. 2022. “Re-Basin via Implicit Sinkhorn Differentiation.” arXiv. https://doi.org/10.48550/ARXIV.2212.12042.
Peng, Baolin, Chunyuan Li, Pengcheng He, Michel Galley, and Jianfeng Gao. 2023. “Instruction Tuning with GPT-4.” arXiv. https://doi.org/10.48550/ARXIV.2304.03277.
Peng, Baolin, Xiujun Li, Lihong Li, Jianfeng Gao, Asli Celikyilmaz, Sungjin Lee, and Kam-Fai Wong. 2017. “Composite Task-Completion Dialogue Policy Learning via Hierarchical Deep Reinforcement Learning.” Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/d17-1237.
Peng, Cheng, Xi Yang, Kaleb E Smith, Zehao Yu, Aokun Chen, Jiang Bian, and Yonghui Wu. 2023. “Model Tuning or Prompt Tuning? A Study of Large Language Models for Clinical Concept and Relation Extraction.” arXiv. https://doi.org/10.48550/ARXIV.2310.06239.
Peng, Huailiang, Yujun Zhang, Hao Sun, Xu Bai, Yangyang Li, and Shuhai Wang. 2022. “Domain-Aware Federated Social Bot Detection with Multi-Relational Graph Neural Networks.” 2022 International Joint Conference on Neural Networks (IJCNN), July. https://doi.org/10.1109/ijcnn55064.2022.9892366.
Peng, Keqin, Liang Ding, Qihuang Zhong, Li Shen, Xuebo Liu, Min Zhang, Yuanxin Ouyang, and Dacheng Tao. 2023. “Towards Making the Most of ChatGPT for Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.2303.13780.
Peng, Peng, Ying Wen, Yaodong Yang, Quan Yuan, Zhenkun Tang, Haitao Long, and Jun Wang. 2017. “Multiagent Bidirectionally-Coordinated Nets: Emergence of Human-Level Coordination in Learning to Play StarCraft Combat Games.” arXiv. https://doi.org/10.48550/ARXIV.1703.10069.
Peng, Shige. 1999. “Monotonic Limit Theorem of BSDE and Nonlinear Decomposition Theorem of Doob–Meyers Type.” Probability Theory and Related Fields 113 (March). https://doi.org/10.1007/s004400050214.
Peng, Songyou, Michael Niemeyer, Lars Mescheder, Marc Pollefeys, and Andreas Geiger. 2020. “Convolutional Occupancy Networks.” arXiv. https://doi.org/10.48550/ARXIV.2003.04618.
Peng, Xiangyu, Kai Wang, Zheng Zhu, Mang Wang, and Yang You. 2022. “Crafting Better Contrastive Views for Siamese Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.2202.03278.
Peng, Yifan, Siddhant Arora, Yosuke Higuchi, Yushi Ueda, Sujay Kumar, Karthik Ganesan, Siddharth Dalmia, Xuankai Chang, and Shinji Watanabe. 2022. “A Study on the Integration of Pre-Trained SSL, ASR, LM and SLU Models for Spoken Language Understanding.” arXiv. https://doi.org/10.48550/ARXIV.2211.05869.
Penha, Gustavo, Alexandru Balan, and Claudia Hauff. 2019. “Introducing MANtIS: A Novel Multi-Domain Information Seeking Dialogues Dataset.” arXiv. https://doi.org/10.48550/ARXIV.1912.04639.
Peri, Neehar, Jonathon Luiten, Mengtian Li, Aljoša Ošep, Laura Leal-Taixé, and Deva Ramanan. 2022. “Forecasting from LiDAR via Future Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2203.16297.
Pesquita, Catia, Daniel Faria, André O. Falcão, Phillip Lord, and Francisco M. Couto. 2009. “Semantic Similarity in Biomedical Ontologies.” PLoS Computational Biology 5 (July). https://doi.org/10.1371/journal.pcbi.1000443.
Peters, Matthew E., Mark Neumann, Luke Zettlemoyer, and Wen-tau Yih. 2018. “Dissecting Contextual Word Embeddings: Architecture and Representation.” arXiv. https://doi.org/10.48550/ARXIV.1808.08949.
Pezzotti, Nicola, Boudewijn P. F. Lelieveldt, Laurens van der Maaten, Thomas Höllt, Elmar Eisemann, and Anna Vilanova. 2015. “Approximated and User Steerable tSNE for Progressive Visual Analytics.” arXiv. https://doi.org/10.48550/ARXIV.1512.01655.
Pfeiffer, Jonas, Andreas Rücklé, Clifton Poth, Aishwarya Kamath, Ivan Vulić, Sebastian Ruder, Kyunghyun Cho, and Iryna Gurevych. 2020. “AdapterHub: A Framework for Adapting Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2007.07779.
Pfeiffer, Jonas, Sebastian Ruder, Ivan Vulić, and Edoardo Maria Ponti. 2023. “Modular Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.2302.11529.
Pfister, Niklas, Evan G. Williams, Jonas Peters, Ruedi Aebersold, and Peter Bühlmann. 2019. “Stabilizing Variable Selection and Regression.” arXiv. https://doi.org/10.48550/ARXIV.1911.01850.
Pham, Hung Viet, Shangshu Qian, Jiannan Wang, Thibaud Lutellier, Jonathan Rosenthal, Lin Tan, Yaoliang Yu, and Nachiappan Nagappan. 2020. “Problems and Opportunities in Training Deep Learning Software Systems.” Proceedings of the 35th IEEE/ACM International Conference on Automated Software Engineering, December. https://doi.org/10.1145/3324884.3416545.
Phillips, P Jonathon, Carina A Hahn, Peter C Fontana, Amy N Yates, Kristen Greene, David A Broniatowski, and Mark A Przybocki. 2021. “Four Principles of Explainable Artificial Intelligence,” September. https://doi.org/10.6028/nist.ir.8312.
Phillips, Steven J., Robert P. Anderson, Miroslav Dudík, Robert E. Schapire, and Mary E. Blair. 2017. “Opening the Black Box: An Open‐source Release of Maxent.” Ecography 40 (May). https://doi.org/10.1111/ecog.03049.
Phipps, Polly, and Daniell Toth. 2012. “Analyzing Establishment Nonresponse Using an Interpretable Regression Tree Model with Linked Administrative Data.” The Annals of Applied Statistics 6 (June). https://doi.org/10.1214/11-aoas521.
Pi, Qi, Weijie Bian, Guorui Zhou, Xiaoqiang Zhu, and Kun Gai. 2019. “Practice on Long Sequential User Behavior Modeling for Click-Through Rate Prediction.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330666.
Picard, Jean. 1996. “On the Existence of Smooth Densities for Jump Processes.” Probability Theory and Related Fields 105 (December). https://doi.org/10.1007/bf01191910.
Pietquin, O., and T. Dutoit. 2006. “A Probabilistic Framework for Dialog Simulation and Optimal Strategy Learning.” IEEE Transactions on Audio, Speech and Language Processing 14 (March). https://doi.org/10.1109/tsa.2005.855836.
Pineau, Joelle, Philippe Vincent-Lamarre, Koustuv Sinha, Vincent Larivière, Alina Beygelzimer, Florence d’Alché-Buc, Emily Fox, and Hugo Larochelle. 2020. “Improving Reproducibility in Machine Learning Research (a Report from the NeurIPS 2019 Reproducibility Program).” arXiv. https://doi.org/10.48550/ARXIV.2003.12206.
Ping, Wei, Kainan Peng, Andrew Gibiansky, Sercan O. Arik, Ajay Kannan, Sharan Narang, Jonathan Raiman, and John Miller. 2017. “Deep Voice 3: Scaling Text-to-Speech with Convolutional Sequence Learning.” arXiv. https://doi.org/10.48550/ARXIV.1710.07654.
Pinto, Lerrel, James Davidson, Rahul Sukthankar, and Abhinav Gupta. 2017. “Robust Adversarial Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1703.02702.
Pipiras, Vladas, and Murad S. Taqqu. 2000. “Integration Questions Related to Fractional Brownian Motion.” Probability Theory and Related Fields 118 (October). https://doi.org/10.1007/s440-000-8016-7.
Piponi, Dan, Dave Moore, and Joshua V. Dillon. 2020. “Joint Distributions for TensorFlow Probability.” arXiv. https://doi.org/10.48550/ARXIV.2001.11819.
Pirrò, Giuseppe. 2020. “Fact Checking via Path Embedding and Aggregation.” arXiv. https://doi.org/10.48550/ARXIV.2011.08028.
Pitman, Jim. 1995. “Exchangeable and Partially Exchangeable Random Partitions.” Probability Theory and Related Fields 102 (June). https://doi.org/10.1007/bf01213386.
Pizzolato, Stefano, Luca Tagliapietra, Matteo Cognolato, Monica Reggiani, Henning Müller, and Manfredo Atzori. 2017. “Comparison of Six Electromyography Acquisition Setups on Hand Movement Classification Tasks.” PLOS ONE 12 (October). https://doi.org/10.1371/journal.pone.0186132.
“Planning for Mobile Manipulation.” 2021, July. https://doi.org/10.25549/USCTHESES-OUC15276668.
Plis, Sergey M., Devon R. Hjelm, Ruslan Salakhutdinov, Elena A. Allen, Henry J. Bockholt, Jeffrey D. Long, Hans J. Johnson, Jane S. Paulsen, Jessica A. Turner, and Vince D. Calhoun. 2014. “Deep Learning for Neuroimaging: A Validation Study.” Frontiers in Neuroscience 8 (August). https://doi.org/10.3389/fnins.2014.00229.
Podinovski, Victor V. 2017. “Returns to Scale in Convex Production Technologies.” European Journal of Operational Research 258 (May). https://doi.org/10.1016/j.ejor.2016.09.029.
Polak, Maciej P., and Dane Morgan. 2023. “Extracting Accurate Materials Data from Research Papers with Conversational Language Models and Prompt Engineering.” arXiv. https://doi.org/10.48550/ARXIV.2303.05352.
Poli, Michael, Stefano Massaroli, Eric Nguyen, Daniel Y. Fu, Tri Dao, Stephen Baccus, Yoshua Bengio, Stefano Ermon, and Christopher Ré. 2023. “Hyena Hierarchy: Towards Larger Convolutional Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.10866.
Polson, Nicholas G., James G. Scott, and Jesse Windle. 2011. “The Bayesian Bridge.” arXiv. https://doi.org/10.48550/ARXIV.1109.2279.
Polzehl, Jörg, and Vladimir Spokoiny. 2005. “Propagation-Separation Approach for Local Likelihood Estimation.” Probability Theory and Related Fields 135 (September). https://doi.org/10.1007/s00440-005-0464-1.
Poria, Soujanya, Erik Cambria, Devamanyu Hazarika, Navonil Majumder, Amir Zadeh, and Louis-Philippe Morency. 2017. “Context-Dependent Sentiment Analysis in User-Generated Videos.” Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/p17-1081.
Portugal, Ivens, Paulo Alencar, and Donald Cowan. 2015. “The Use of Machine Learning Algorithms in Recommender Systems: A Systematic Review.” arXiv. https://doi.org/10.48550/ARXIV.1511.05263.
Pourkamali, Nooshin, and Shler Ebrahim Sharifi. 2024. “Machine Translation with Large Language Models: Prompt Engineering for Persian, English, and Russian Directions.” arXiv. https://doi.org/10.48550/ARXIV.2401.08429.
Pourreza, Mohammadreza, and Davood Rafiei. 2023. “DIN-SQL: Decomposed in-Context Learning of Text-to-SQL with Self-Correction.” arXiv. https://doi.org/10.48550/ARXIV.2304.11015.
Pourzanjani, Arya A., Richard M. Jiang, Brian Mitchell, Paul J. Atzberger, and Linda R. Petzold. 2021. “Bayesian Inference over the Stiefel Manifold via the Givens Representation.” Bayesian Analysis 16 (June). https://doi.org/10.1214/20-ba1202.
Poyarkov, Alexey, Alexey Drutsa, Andrey Khalyavin, Gleb Gusev, and Pavel Serdyukov. 2016. “Boosted Decision Tree Regression Adjustment for Variance Reduction in Online Controlled Experiments.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939688.
Prabhu, Ameya, Vishal Batchu, Sri Aurobindo Munagala, Rohit Gajawada, and Anoop Namboodiri. 2018. “Distribution-Aware Binarization of Neural Networks for Sketch Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1804.02941.
Prabhu, Ameya, Hasan Abed Al Kader Hammoud, Puneet Dokania, Philip H. S. Torr, Ser-Nam Lim, Bernard Ghanem, and Adel Bibi. 2023. “Computationally Budgeted Continual Learning: What Does Matter?” arXiv. https://doi.org/10.48550/ARXIV.2303.11165.
Prado, Miguel de, Maurizio Denna, Luca Benini, and Nuria Pazos. 2018. “QUENN.” Proceedings of the 15th ACM International Conference on Computing Frontiers, May. https://doi.org/10.1145/3203217.3203282.
Prantl, Lukas, Nuttapong Chentanez, Stefan Jeschke, and Nils Thuerey. 2019. “Tranquil Clouds: Neural Networks for Learning Temporally Coherent Features in Point Clouds.” arXiv. https://doi.org/10.48550/ARXIV.1907.05279.
Prather, James, Paul Denny, Juho Leinonen, David H. Smith, Brent N. Reeves, Stephen MacNeil, Brett A. Becker, Andrew Luxton-Reilly, Thezyrie Amarouche, and Bailey Kimmel. 2024. “Interactions with Prompt Problems: A New Way to Teach Programming with Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.10759.
Predd, J. B., S. B. Kulkarni, and H. V. Poor. 2006. “Distributed Learning in Wireless Sensor Networks.” IEEE Signal Processing Magazine 23 (July). https://doi.org/10.1109/msp.2006.1657817.
Price, Robert, Rob Bethune, and Lisa Massey. 2019. “Problem with p Values: Why p Values Do Not Tell You If Your Treatment Is Likely to Work.” Postgraduate Medical Journal 96 (October). https://doi.org/10.1136/postgradmedj-2019-137079.
Pritikin, Joshua N., Lance M. Rappaport, and Michael C. Neale. 2017. “Likelihood-Based Confidence Intervals for a Parameter with an Upper or Lower Bound.” Structural Equation Modeling: A Multidisciplinary Journal 24 (January). https://doi.org/10.1080/10705511.2016.1275969.
“Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.” 2007, August. https://doi.org/10.1145/1281192.
“Proceedings of the 1st ACM SIGPLAN International Workshop on Machine Learning and Programming Languages.” 2017, June. https://doi.org/10.1145/3088525.
“Proceedings of the 21st ACM SIGPLAN International Conference on Functional Programming.” 2016, September. https://doi.org/10.1145/2951913.
“Proceedings of the 23rd International Conference on Machine Learning - ICML ’06.” 2006. https://doi.org/10.1145/1143844.
“Proceedings of the 2nd French-Speaking Conference on Mobility and Uibquity Computing - UbiMob ’05.” 2005. https://doi.org/10.1145/1102613.
“Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.” 2020. https://doi.org/10.18653/v1/2020.acl-main.
“Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence.” 2019, August. https://doi.org/10.24963/ijcai.2019.
“Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence.” 2018, July. https://doi.org/10.24963/ijcai.2018.
“Proceedings of the Web Conference 2021.” 2021, April. https://doi.org/10.1145/3442381.
“Proceedings of Third International Conference on Sustainable Expert Systems.” 2023. Lecture Notes in Networks and Systems. https://doi.org/10.1007/978-981-19-7874-6.
“Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications.” 2010. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-642-16687-7.
Prokudin, Sergey, Christoph Lassner, and Javier Romero. 2019. “Efficient Learning on Point Clouds with Basis Point Sets.” arXiv. https://doi.org/10.48550/ARXIV.1908.09186.
Provost, Foster, and Tom Fawcett. 2000. “Robust Classification for Imprecise Environments.” arXiv. https://doi.org/10.48550/ARXIV.CS/0009007.
Pryzant, Reid, Dan Iter, Jerry Li, Yin Tat Lee, Chenguang Zhu, and Michael Zeng. 2023. “Automatic Prompt Optimization with "Gradient Descent" and Beam Search.” arXiv. https://doi.org/10.48550/ARXIV.2305.03495.
Pugh, Justin K., Lisa B. Soros, and Kenneth O. Stanley. 2016. “Quality Diversity: A New Frontier for Evolutionary Computation.” Frontiers in Robotics and AI 3 (July). https://doi.org/10.3389/frobt.2016.00040.
Puigcerver, Joan. 2017. “Are Multidimensional Recurrent Layers Really Necessary for Handwritten Text Recognition?” 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), November. https://doi.org/10.1109/icdar.2017.20.
Punjabi, Surabhi, and Priyanka Bhatt. 2018. “Robust Factorization Machines for User Response Prediction.” Proceedings of the 2018 World Wide Web Conference on World Wide Web - WWW ’18. https://doi.org/10.1145/3178876.3186148.
Purdom, Elizabeth. 2011. “Analysis of a Data Matrix and a Graph: Metagenomic Data and the Phylogenetic Tree.” The Annals of Applied Statistics 5 (December). https://doi.org/10.1214/10-aoas402.
Putter, F. A. M. de, and Henk Corporaal. 2023. “How to Train Accurate BNNs for Embedded Systems?” Embedded Machine Learning for Cyber-Physical, IoT, and Edge Computing, October. https://doi.org/10.1007/978-3-031-39932-9_5.
Pyysalo, Sampo, Antti Airola, Juho Heimonen, Jari Björne, Filip Ginter, and Tapio Salakoski. 2008. “Comparative Analysis of Five Protein-Protein Interaction Corpora.” BMC Bioinformatics 9 (April). https://doi.org/10.1186/1471-2105-9-s3-s6.
Qayyum, Adnan, Junaid Qadir, Muhammad Bilal, and Ala Al-Fuqaha. 2020. “Secure and Robust Machine Learning for Healthcare: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2001.08103.
Qi, Charles R., Hao Su, Kaichun Mo, and Leonidas J. Guibas. 2016. “PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1612.00593.
Qi, Fanchao, Chenghao Yang, Zhiyuan Liu, Qiang Dong, Maosong Sun, and Zhendong Dong. 2019. “OpenHowNet: An Open Sememe-Based Lexical Knowledge Base.” arXiv. https://doi.org/10.48550/ARXIV.1901.09957.
Qi, Panpan, Wei Wang, Lei Zhu, and See Kiong Ng. 2021. “Unsupervised Domain Adaptation for Static Malware Detection Based on Gradient Boosting Trees.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3482400.
Qi, Pi, Xiaoqiang Zhu, Guorui Zhou, Yujing Zhang, Zhe Wang, Lejian Ren, Ying Fan, and Kun Gai. 2020. “Search-Based User Interest Modeling with Lifelong Sequential Behavior Data for Click-Through Rate Prediction,” June. http://arxiv.org/abs/2006.05639v2.
Qian, Biao, and Yang Wang. 2019. “A Targeted Acceleration and Compression Framework for Low Bit Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1907.05271.
Qian, Chen, Huayi Tang, Zhirui Yang, Hong Liang, and Yong Liu. 2023. “Can Large Language Models Empower Molecular Property Prediction?” arXiv. https://doi.org/10.48550/ARXIV.2307.07443.
Qian, Yujie, Enrico Santus, Zhijing Jin, Jiang Guo, and Regina Barzilay. 2018. “GraphIE: A Graph-Based Framework for Information Extraction.” arXiv. https://doi.org/10.48550/ARXIV.1810.13083.
Qin, Can, Haoxuan You, Lichen Wang, C. -C. Jay Kuo, and Yun Fu. 2019. “PointDAN: A Multi-Scale 3D Domain Adaption Network for Point Cloud Representation.” arXiv. https://doi.org/10.48550/ARXIV.1911.02744.
Qin, Chengwei, and Shafiq Joty. 2021. “LFPT5: A Unified Framework for Lifelong Few-Shot Language Learning Based on Prompt Tuning of T5.” arXiv. https://doi.org/10.48550/ARXIV.2110.07298.
Qin, Tao, and Tie-Yan Liu. 2013. “Introducing LETOR 4.0 Datasets.” arXiv. https://doi.org/10.48550/ARXIV.1306.2597.
Qin, Tian, Alex Beatson, Deniz Oktay, Nick McGreivy, and Ryan P. Adams. 2022. “Meta-PDE: Learning to Solve PDEs Quickly Without a Mesh.” arXiv. https://doi.org/10.48550/ARXIV.2211.01604.
Qin, Yao, Dongjin Song, Haifeng Chen, Wei Cheng, Guofei Jiang, and Garrison Cottrell. 2017. “A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1704.02971.
Qin, Yujia, Shengding Hu, Yankai Lin, Weize Chen, Ning Ding, Ganqu Cui, Zheni Zeng, et al. 2023. “Tool Learning with Foundation Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.08354.
Qin, Yujia, Shihao Liang, Yining Ye, Kunlun Zhu, Lan Yan, Yaxi Lu, Yankai Lin, et al. 2023. “ToolLLM: Facilitating Large Language Models to Master 16000+ Real-World APIs.” arXiv. https://doi.org/10.48550/ARXIV.2307.16789.
Qin, Zhuwei, Fuxun Yu, ChenChen Liu, and Xiang Chen. 2018. “Demystifying Neural Network Filter Pruning.” arXiv. https://doi.org/10.48550/ARXIV.1811.02639.
Qin, Zidi, Yang Liu, Qing He, and Xiang Ao. 2022. “Explainable Graph-Based Fraud Detection via Neural Meta-Graph Search.” Proceedings of the 31st ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3511808.3557598.
Qiu, David, Yanzhang He, Qiujia Li, Yu Zhang, Liangliang Cao, and Ian McGraw. 2021. “Multi-Task Learning for End-to-End ASR Word and Utterance Confidence with Deletion Prediction.” arXiv. https://doi.org/10.48550/ARXIV.2104.12870.
Qiu, Jiezhong, Qibin Chen, Yuxiao Dong, Jing Zhang, Hongxia Yang, Ming Ding, Kuansan Wang, and Jie Tang. 2020. “GCC.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403168.
Qiu, Jiezhong, Jian Tang, Hao Ma, Yuxiao Dong, Kuansan Wang, and Jie Tang. 2018. “DeepInf.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220077.
Qiu, Xiafei, Wubin Cen, Zhengping Qian, You Peng, Ying Zhang, Xuemin Lin, and Jingren Zhou. 2018. “Real-Time Constrained Cycle Detection in Large Dynamic Graphs.” Proceedings of the VLDB Endowment 11 (August). https://doi.org/10.14778/3229863.3229874.
Qiu, XiPeng, TianXiang Sun, YiGe Xu, YunFan Shao, Ning Dai, and XuanJing Huang. 2020. “Pre-Trained Models for Natural Language Processing: A Survey.” Science China Technological Sciences 63 (September). https://doi.org/10.1007/s11431-020-1647-3.
Qu, Yanru, Han Cai, Kan Ren, Weinan Zhang, Yong Yu, Ying Wen, and Jun Wang. 2016. “Product-Based Neural Networks for User Response Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1611.00144.
Qu, Zhe, Xingyu Li, Xiao Han, Rui Duan, Chengchao Shen, and Lixing Chen. 2023. “How to Prevent the Poor Performance Clients for Personalized Federated Learning?” 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52729.2023.01171.
Quadrana, Massimo, Paolo Cremonesi, and Dietmar Jannach. 2018. “Sequence-Aware Recommender Systems.” arXiv. https://doi.org/10.48550/ARXIV.1802.08452.
Quadrana, Massimo, Antoine Larreche-Mouly, and Matthias Mauch. 2022. “Multi-Objective Hyper-Parameter Optimization of Behavioral Song Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.2208.12724.
Quiroz, Matias, David J. Nott, and Robert Kohn. 2023. “Gaussian Variational Approximations for High-Dimensional State Space Models.” Bayesian Analysis 18 (September). https://doi.org/10.1214/22-ba1332.
Rabin, Md Rafiqul Islam, Aftab Hussain, Mohammad Amin Alipour, and Vincent J. Hellendoorn. 2023. “Memorization and Generalization in Neural Code Intelligence Models.” Information and Software Technology 153 (January). https://doi.org/10.1016/j.infsof.2022.107066.
Rackauckas, Christopher, Yingbo Ma, Julius Martensen, Collin Warner, Kirill Zubov, Rohit Supekar, Dominic Skinner, Ali Ramadhan, and Alan Edelman. 2020. “Universal Differential Equations for Scientific Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.2001.04385.
Rackauckas, Christopher, and Qing Nie. 2017. “DifferentialEquations.jl – a Performant and Feature-Rich Ecosystem for Solving Differential Equations in Julia.” Journal of Open Research Software 5 (May). https://doi.org/10.5334/jors.151.
Rad, M. Torabi, A. Viardin, G. J. Schmitz, and M. Apel. 2019. “Theory-Training Deep Neural Networks for an Alloy Solidification Benchmark Problem.” arXiv. https://doi.org/10.48550/ARXIV.1912.09800.
Radford, Alec, Rafal Jozefowicz, and Ilya Sutskever. 2017. “Learning to Generate Reviews and Discovering Sentiment.” arXiv. https://doi.org/10.48550/ARXIV.1704.01444.
Radlinski, Filip, and Nick Craswell. 2013. “Optimized Interleaving for Online Retrieval Evaluation.” Proceedings of the Sixth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/2433396.2433429.
Radovanovic, Milos, Alexandros Nanopoulos, and Mirjana Ivanovic. 2015. “Reverse Nearest Neighbors in Unsupervised Distance-Based Outlier Detection.” IEEE Transactions on Knowledge and Data Engineering 27 (May). https://doi.org/10.1109/tkde.2014.2365790.
Rafailov, Rafael, Archit Sharma, Eric Mitchell, Stefano Ermon, Christopher D. Manning, and Chelsea Finn. 2023. “Direct Preference Optimization: Your Language Model Is Secretly a Reward Model.” arXiv. https://doi.org/10.48550/ARXIV.2305.18290.
Rahal, Najoua, Maroua Tounsi, and Adel M. Alimi. 2018. “Neural Probabilistic System for Text Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1812.03680.
Rahimi, Ali, and Benjamin Recht. 2008. “Uniform Approximation of Functions with Random Bases.” 2008 46th Annual Allerton Conference on Communication, Control, and Computing, September. https://doi.org/10.1109/allerton.2008.4797607.
Rahman, Md Musfiqur, and Murat Kocaoglu. 2024. “Modular Learning of Deep Causal Generative Models for High-Dimensional Causal Inference.” arXiv. https://doi.org/10.48550/ARXIV.2401.01426.
Rahman, Md. Mostafizer, and Yutaka Watanobe. 2023. “ChatGPT for Education and Research: Opportunities, Threats, and Strategies.” Applied Sciences 13 (May). https://doi.org/10.3390/app13095783.
Raissi, Maziar, Alireza Yazdani, and George Em Karniadakis. 2020. “Hidden Fluid Mechanics: Learning Velocity and Pressure Fields from Flow Visualizations.” Science 367 (February). https://doi.org/10.1126/science.aaw4741.
Raissi, M., P. Perdikaris, and G. E. Karniadakis. 2019. “Physics-Informed Neural Networks: A Deep Learning Framework for Solving Forward and Inverse Problems Involving Nonlinear Partial Differential Equations.” Journal of Computational Physics 378 (February). https://doi.org/10.1016/j.jcp.2018.10.045.
Rajani, Nazneen Fatema, Bryan McCann, Caiming Xiong, and Richard Socher. 2019. “Explain Yourself! Leveraging Language Models for Commonsense Reasoning.” arXiv. https://doi.org/10.48550/ARXIV.1906.02361.
Rajapakse, Dilina Chandika, and Douglas Leith. 2022. “Fast and Accurate User Cold-Start Learning Using Monte Carlo Tree Search.” Proceedings of the 16th ACM Conference on Recommender Systems, September. https://doi.org/10.1145/3523227.3546786.
Rajkomar, Alvin, Eyal Oren, Kai Chen, Andrew M. Dai, Nissan Hajaj, Michaela Hardt, Peter J. Liu, et al. 2018. “Scalable and Accurate Deep Learning with Electronic Health Records.” Npj Digital Medicine 1 (May). https://doi.org/10.1038/s41746-018-0029-1.
Rajpurkar, Pranav, Jian Zhang, Konstantin Lopyrev, and Percy Liang. 2016. “SQuAD: 100,000+ Questions for Machine Comprehension of Text.” arXiv. https://doi.org/10.48550/ARXIV.1606.05250.
Rakhshani, Hojjat, Hassan Ismail Fawaz, Lhassane Idoumghar, Germain Forestier, Julien Lepagnot, Jonathan Weber, Mathieu Brevilliers, and Pierre-Alain Muller. 2020. “Neural Architecture Search for Time Series Classification.” 2020 International Joint Conference on Neural Networks (IJCNN), July. https://doi.org/10.1109/ijcnn48605.2020.9206721.
Ramachandran, Prajit, Peter J. Liu, and Quoc V. Le. 2016. “Unsupervised Pretraining for Sequence to Sequence Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.02683.
Ramakrishnan, Ramchalam Kinattinkara, Eyyüb Sari, and Vahid Partovi Nia. 2019. “Differentiable Mask for Pruning Convolutional and Recurrent Networks.” arXiv. https://doi.org/10.48550/ARXIV.1909.04567.
Ramanath, Rohan, Hakan Inan, Gungor Polatkan, Bo Hu, Qi Guo, Cagri Ozcaglar, Xianren Wu, Krishnaram Kenthapadi, and Sahin Cem Geyik. 2018. “Towards Deep and Representation Learning for Talent Search at LinkedIn.” arXiv. https://doi.org/10.48550/ARXIV.1809.06473.
Ramasinghe, Sameera, Salman Khan, Nick Barnes, and Stephen Gould. 2019. “Spectral-GANs for High-Resolution 3D Point-Cloud Generation.” arXiv. https://doi.org/10.48550/ARXIV.1912.01800.
Ranganath, Rajesh, Sean Gerrish, and David M. Blei. 2014. “Black Box Variational Inference.” arXiv. https://doi.org/10.48550/ARXIV.1401.0118.
Rao, Arya, John Kim, Meghana Kamineni, Michael Pang, Winston Lie, and Marc D. Succi. 2023. “Evaluating ChatGPT as an Adjunct for Radiologic Decision-Making,” February. https://doi.org/10.1101/2023.02.02.23285399.
Rao, Arya, Michael Pang, John Kim, Meghana Kamineni, Winston Lie, Anoop K. Prasad, Adam Landman, Keith J Dreyer, and Marc D. Succi. 2023. “Assessing the Utility of ChatGPT Throughout the Entire Clinical Workflow,” February. https://doi.org/10.1101/2023.02.21.23285886.
Rao, Susie Xi, Shuai Zhang, Zhichao Han, Zitao Zhang, Wei Min, Zhiyao Chen, Yinan Shan, Yang Zhao, and Ce Zhang. 2021. “xFraud.” Proceedings of the VLDB Endowment 15 (November). https://doi.org/10.14778/3494124.3494128.
Raposo, David, Adam Santoro, David Barrett, Razvan Pascanu, Timothy Lillicrap, and Peter Battaglia. 2017. “Discovering Objects and Their Relations from Entangled Scene Representations.” arXiv. https://doi.org/10.48550/ARXIV.1702.05068.
Rasmussen, Jens. 1985. “The Role of Hierarchical Knowledge Representation in Decisionmaking and System Management.” IEEE Transactions on Systems, Man, and Cybernetics SMC-15 (March). https://doi.org/10.1109/tsmc.1985.6313353.
Rastegari, Mohammad, Vicente Ordonez, Joseph Redmon, and Ali Farhadi. 2016. “XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1603.05279.
Rasul, Kashif, Abdul-Saboor Sheikh, Ingmar Schuster, Urs Bergmann, and Roland Vollgraf. 2020. “Multivariate Probabilistic Time Series Forecasting via Conditioned Normalizing Flows.” arXiv. https://doi.org/10.48550/ARXIV.2002.06103.
Ratner, Alexander, Stephen H. Bach, Henry Ehrenberg, Jason Fries, Sen Wu, and Christopher Ré. 2017. “Snorkel.” Proceedings of the VLDB Endowment 11 (November). https://doi.org/10.14778/3157794.3157797.
Räuker, Tilman, Anson Ho, Stephen Casper, and Dylan Hadfield-Menell. 2022. “Toward Transparent AI: A Survey on Interpreting the Inner Structures of Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2207.13243.
Ravanbakhsh, Siamak, Jeff Schneider, and Barnabas Poczos. 2016. “Deep Learning with Sets and Point Clouds.” arXiv. https://doi.org/10.48550/ARXIV.1611.04500.
Ravanelli, Mirco, Philemon Brakel, Maurizio Omologo, and Yoshua Bengio. 2017. “A Network of Deep Neural Networks for Distant Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1703.08002.
Ravikumar, Pradeep, Martin J. Wainwright, Garvesh Raskutti, and Bin Yu. 2008. “High-Dimensional Covariance Estimation by Minimizing \(\ell_1\)-Penalized Log-Determinant Divergence.” arXiv. https://doi.org/10.48550/ARXIV.0811.3628.
Rawte, Vipula, Prachi Priya, S. M Towhidul Islam Tonmoy, S M Mehedi Zaman, Amit Sheth, and Amitava Das. 2023. “Exploring the Relationship Between LLM Hallucinations and Prompt Linguistic Nuances: Readability, Formality, and Concreteness.” arXiv. https://doi.org/10.48550/ARXIV.2309.11064.
Rayana, Shebuti, and Leman Akoglu. 2015. “Collective Opinion Spam Detection.” Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2783258.2783370.
Razavi, Ali, Aaron van den Oord, and Oriol Vinyals. 2019. “Generating Diverse High-Fidelity Images with VQ-VAE-2.” arXiv. https://doi.org/10.48550/ARXIV.1906.00446.
Razdaibiedina, Anastasia, Yuning Mao, Rui Hou, Madian Khabsa, Mike Lewis, and Amjad Almahairi. 2023. “Progressive Prompts: Continual Learning for Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2301.12314.
Reagen, Brandon, Udit Gupta, Robert Adolf, Michael M. Mitzenmacher, Alexander M. Rush, Gu-Yeon Wei, and David Brooks. 2017. “Weightless: Lossy Weight Encoding for Deep Neural Network Compression.” arXiv. https://doi.org/10.48550/ARXIV.1711.04686.
Real, Esteban, Alok Aggarwal, Yanping Huang, and Quoc V Le. 2018. “Regularized Evolution for Image Classifier Architecture Search.” arXiv. https://doi.org/10.48550/ARXIV.1802.01548.
“Recent Advances in Reinforcement Learning.” 2012. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-642-29946-9.
Recht, Benjamin, and Christopher Re. 2012. “Beneath the Valley of the Noncommutative Arithmetic-Geometric Mean Inequality: Conjectures, Case-Studies, and Consequences.” arXiv. https://doi.org/10.48550/ARXIV.1202.4184.
Reddi, Vijay Janapa, David Kanter, Peter Mattson, Jared Duke, Thai Nguyen, Ramesh Chukka, Ken Shiring, et al. 2020. “MLPerf Mobile Inference Benchmark.” arXiv. https://doi.org/10.48550/ARXIV.2012.02328.
Redell, Nickalus. 2019. “Shapley Decomposition of r-Squared in Machine Learning Models.” arXiv. https://doi.org/10.48550/ARXIV.1908.09718.
Redmon, Joseph, and Ali Farhadi. 2016. “YOLO9000: Better, Faster, Stronger.” arXiv. https://doi.org/10.48550/ARXIV.1612.08242.
Reece, Andrew G., and Christopher M. Danforth. 2016. “Instagram Photos Reveal Predictive Markers of Depression.” arXiv. https://doi.org/10.48550/ARXIV.1608.03282.
Reed, Colorado J, Sean Metzger, Aravind Srinivas, Trevor Darrell, and Kurt Keutzer. 2020. “SelfAugment: Automatic Augmentation Policies for Self-Supervised Learning.” arXiv. https://doi.org/10.48550/ARXIV.2009.07724.
Reed, Scott, Zeynep Akata, Xinchen Yan, Lajanugen Logeswaran, Bernt Schiele, and Honglak Lee. 2016. “Generative Adversarial Text to Image Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1605.05396.
Reed, Scott, Konrad Zolna, Emilio Parisotto, Sergio Gomez Colmenarejo, Alexander Novikov, Gabriel Barth-Maron, Mai Gimenez, et al. 2022. “A Generalist Agent.” arXiv. https://doi.org/10.48550/ARXIV.2205.06175.
Reiter-Haas, Markus, Emanuel Lacic, Tomislav Duricic, Valentin Slawicek, and Elisabeth Lex. 2019. “Should We Embed? A Study on the Online Performance of Utilizing Embeddings for Real-Time Job Recommendations.” arXiv. https://doi.org/10.48550/ARXIV.1907.06556.
Ren, Hansheng, Bixiong Xu, Yujing Wang, Chao Yi, Congrui Huang, Xiaoyu Kou, Tony Xing, Mao Yang, Jie Tong, and Qi Zhang. 2019. “Time-Series Anomaly Detection Service at Microsoft.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330680.
Ren, Kan, Weinan Zhang, Yifei Rong, Haifeng Zhang, Yong Yu, and Jun Wang. 2016. “User Response Learning for Directly Optimizing Campaign Performance in Display Advertising.” Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, October. https://doi.org/10.1145/2983323.2983347.
Ren, Mengye, Eleni Triantafillou, Sachin Ravi, Jake Snell, Kevin Swersky, Joshua B. Tenenbaum, Hugo Larochelle, and Richard S. Zemel. 2018. “Meta-Learning for Semi-Supervised Few-Shot Classification.” arXiv. https://doi.org/10.48550/ARXIV.1803.00676.
Ren, Shaoqing, Kaiming He, Ross Girshick, and Jian Sun. 2015. “Faster r-CNN: Towards Real-Time Object Detection with Region Proposal Networks.” arXiv. https://doi.org/10.48550/ARXIV.1506.01497.
Ren, Sucheng, Daquan Zhou, Shengfeng He, Jiashi Feng, and Xinchao Wang. 2021. “Shunted Self-Attention via Multi-Scale Token Aggregation.” arXiv. https://doi.org/10.48550/ARXIV.2111.15193.
Ren, Xiaozhe, Pingyi Zhou, Xinfan Meng, Xinjing Huang, Yadao Wang, Weichao Wang, Pengfei Li, et al. 2023. “PanGu-σ: Towards Trillion Parameter Language Model with Sparse Heterogeneous Computing.” arXiv. https://doi.org/10.48550/ARXIV.2303.10845.
Ren, Yafeng, and Donghong Ji. 2017. “Neural Networks for Deceptive Opinion Spam Detection: An Empirical Study.” Information Sciences 385-386 (April). https://doi.org/10.1016/j.ins.2017.01.015.
Ren, Yi, Yangjun Ruan, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, and Tie-Yan Liu. 2019. “FastSpeech: Fast, Robust and Controllable Text to Speech.” arXiv. https://doi.org/10.48550/ARXIV.1905.09263.
Ren, Yuxiang, Bo Wang, Jiawei Zhang, and Yi Chang. 2021. “Adversarial Active Learning Based Heterogeneous Graph Neural Network for Fake News Detection.” arXiv. https://doi.org/10.48550/ARXIV.2101.11206.
Rendle, Steffen, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. 2012. “BPR: Bayesian Personalized Ranking from Implicit Feedback.” arXiv. https://doi.org/10.48550/ARXIV.1205.2618.
Rethmeier, Nils, and Isabelle Augenstein. 2021. “A Primer on Contrastive Pretraining in Language Processing: Methods, Lessons Learned and Perspectives.” arXiv. https://doi.org/10.48550/ARXIV.2102.12982.
Reyzin, Lev, and Robert E. Schapire. 2006. “How Boosting the Margin Can Also Boost Classifier Complexity.” Proceedings of the 23rd International Conference on Machine Learning - ICML ’06. https://doi.org/10.1145/1143844.1143939.
Rezayi, Saed, Handong Zhao, Sungchul Kim, Ryan Rossi, Nedim Lipka, and Sheng Li. 2021. “Edge: Enriching Knowledge Graph Embeddings with External Text.” Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. https://doi.org/10.18653/v1/2021.naacl-main.221.
Rezende, Danilo Jimenez, Shakir Mohamed, and Daan Wierstra. 2014. “Stochastic Backpropagation and Approximate Inference in Deep Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.1401.4082.
Rhee, Sungmin, Seokjun Seo, and Sun Kim. 2018. “Hybrid Approach of Relation Network and Localized Graph Convolutional Filtering for Breast Cancer Subtype Classification.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/490.
Ribeiro, Leonardo F. R., Pedro H. P. Saverese, and Daniel R. Figueiredo. 2017. “<I>struc2vec</i>.” Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3097983.3098061.
Ribeiro, Marco Tulio, Sameer Singh, and Carlos Guestrin. 2016a. “Model-Agnostic Interpretability of Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.05386.
———. 2016b. “"Why Should i Trust You?".” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939778.
Ribeiro, Rui, Joao P. Carvalho, and Luísa Coheur. 2023. “PGTask: Introducing the Task of Profile Generation from Dialogues.” arXiv. https://doi.org/10.48550/ARXIV.2304.06634.
Rice, Leslie, Eric Wong, and J. Zico Kolter. 2020. “Overfitting in Adversarially Robust Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.2002.11569.
Richardson, Sylvia., and Peter J. Green. 1997. “On Bayesian Analysis of Mixtures with an Unknown Number of Components (with Discussion).” Journal of the Royal Statistical Society Series B: Statistical Methodology 59 (November). https://doi.org/10.1111/1467-9868.00095.
Richtárik, Peter, and Martin Takáč. 2012. “Parallel Coordinate Descent Methods for Big Data Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1212.0873.
Riesenhuber, Maximilian, and Tomaso Poggio. 1999. “Hierarchical Models of Object Recognition in Cortex.” Nature Neuroscience 2 (November). https://doi.org/10.1038/14819.
Rissanen, J., and G. Langdon. 1981. “Universal Modeling and Coding.” IEEE Transactions on Information Theory 27 (January). https://doi.org/10.1109/tit.1981.1056282.
Rivas, Pablo, and Liang Zhao. 2023. “Marketing with ChatGPT: Navigating the Ethical Terrain of GPT-Based Chatbot Technology.” AI 4 (April). https://doi.org/10.3390/ai4020019.
Rivest, R. L., A. Shamir, and L. Adleman. 1978. “A Method for Obtaining Digital Signatures and Public-Key Cryptosystems.” Communications of the ACM 21 (February). https://doi.org/10.1145/359340.359342.
Rivolli, Adriano, Luís P. F. Garcia, Carlos Soares, Joaquin Vanschoren, and André C. P. L. F. de Carvalho. 2018. “Characterizing Classification Datasets: A Study of Meta-Features for Meta-Learning.” arXiv. https://doi.org/10.48550/ARXIV.1808.10406.
Rizve, Mamshad Nayeem, Kevin Duarte, Yogesh S Rawat, and Mubarak Shah. 2021. “In Defense of Pseudo-Labeling: An Uncertainty-Aware Pseudo-Label Selection Framework for Semi-Supervised Learning.” arXiv. https://doi.org/10.48550/ARXIV.2101.06329.
Rizve, Mamshad Nayeem, Navid Kardan, and Mubarak Shah. 2022. “Towards Realistic Semi-Supervised Learning.” arXiv. https://doi.org/10.48550/ARXIV.2207.02269.
Robert, Christian P., and Wu Changye. 2020. “Markov Chain Monte Carlo Methods, a Survey with Some Frequent Misunderstandings.” arXiv. https://doi.org/10.48550/ARXIV.2001.06249.
Robles, J. Gomez, and J. Vanschoren. 2019. “Learning to Reinforcement Learn for Neural Architecture Search.” arXiv. https://doi.org/10.48550/ARXIV.1911.03769.
Rodriguez, Manuel Gomez, David Balduzzi, and Bernhard Schölkopf. 2011. “Uncovering the Temporal Dynamics of Diffusion Networks.” arXiv. https://doi.org/10.48550/ARXIV.1105.0697.
Roffe, Joschka. 2019. “Quantum Error Correction: An Introductory Guide.” Contemporary Physics 60 (July). https://doi.org/10.1080/00107514.2019.1667078.
Roh, Byungseok, Wuhyun Shin, Ildoo Kim, and Sungwoong Kim. 2021. “Spatially Consistent Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.2103.06122.
Rohanian, Omid, Mohammadmahdi Nouriborji, Samaneh Kouchaki, and David A Clifton. 2023. “On the Effectiveness of Compact Biomedical Transformers.” Bioinformatics 39 (February). https://doi.org/10.1093/bioinformatics/btad103.
Roller, Stephen, Emily Dinan, Naman Goyal, Da Ju, Mary Williamson, Yinhan Liu, Jing Xu, et al. 2020. “Recipes for Building an Open-Domain Chatbot.” arXiv. https://doi.org/10.48550/ARXIV.2004.13637.
Roller, Stephen, Emily Dinan, Naman Goyal, Da Ju, Mary Williamson, Yinhan Liu, Jing Xu, et al. 2021. “Recipes for Building an Open-Domain Chatbot.” Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume. https://doi.org/10.18653/v1/2021.eacl-main.24.
Romero, Adriana, Nicolas Ballas, Samira Ebrahimi Kahou, Antoine Chassang, Carlo Gatta, and Yoshua Bengio. 2014. “FitNets: Hints for Thin Deep Nets.” arXiv. https://doi.org/10.48550/ARXIV.1412.6550.
Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. 2015. “U-Net: Convolutional Networks for Biomedical Image Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1505.04597.
RONQUIST, F. 2004. “Bayesian Inference of Character Evolution.” Trends in Ecology &Amp; Evolution 19 (September). https://doi.org/10.1016/j.tree.2004.07.002.
Ronquist, Fredrik, Maxim Teslenko, Paul van der Mark, Daniel L. Ayres, Aaron Darling, Sebastian Höhna, Bret Larget, Liang Liu, Marc A. Suchard, and John P. Huelsenbeck. 2012. “MrBayes 3.2: Efficient Bayesian Phylogenetic Inference and Model Choice Across a Large Model Space.” Systematic Biology 61 (February). https://doi.org/10.1093/sysbio/sys029.
Roohani, Yusuf, Kexin Huang, and Jure Leskovec. 2023. “Predicting Transcriptional Outcomes of Novel Multigene Perturbations with GEARS.” Nature Biotechnology, August. https://doi.org/10.1038/s41587-023-01905-6.
Rosenman, Evan, Santiago Olivella, and Kosuke Imai. 2022. “Name Dictionaries for "Wru" r Package.” https://doi.org/10.7910/DVN/7TRYAC.
Ross, Cody T., Richard McElreath, and Daniel Redhead. 2022. “Modelling Human and Non-Human Animal Network Data in r Using STRAND,” May. https://doi.org/10.1101/2022.05.13.491798.
———. 2023. “Modelling Animal Network Data in r Using <Scp>STRAND</Scp>.” Journal of Animal Ecology, November. https://doi.org/10.1111/1365-2656.14021.
Rossi, V., and J.-P. Vila. 2006. “Bayesian Multioutput Feedforward Neural Networks Comparison: A Conjugate Prior Approach.” IEEE Transactions on Neural Networks 17 (January). https://doi.org/10.1109/tnn.2005.860883.
Roth, Holger, Le Lu, Ari Seff, Kevin M Cherry, Joanne Hoffman, Shijun Wang, Jiamin Liu, Evrim Turkbey, and Ronald M. Summers. 2015. “A New 2.5 d Representation for Lymph Node Detection in CT (CT Lymph Nodes).” https://doi.org/10.7937/K9/TCIA.2015.AQIIDCNM.
Rothe, Sascha, Shashi Narayan, and Aliaksei Severyn. 2019. “Leveraging Pre-Trained Checkpoints for Sequence Generation Tasks.” arXiv. https://doi.org/10.48550/ARXIV.1907.12461.
Rouhani, Bita Darvish, Huili Chen, and Farinaz Koushanfar. 2018. “DeepSigns: A Generic Watermarking Framework for IP Protection of Deep Learning Models.” arXiv. https://doi.org/10.48550/ARXIV.1804.00750.
Roumeliotis, Konstantinos I., Nikolaos D. Tselikas, and Dimitrios K. Nasiopoulos. 2023. “Llama 2: Early Adopters’ Utilization of Meta’s New Open-Source Pretrained Model,” August. https://doi.org/10.20944/preprints202307.2142.v2.
Roux, Nicolas Le, Mark Schmidt, and Francis Bach. 2012. “A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets.” arXiv. https://doi.org/10.48550/ARXIV.1202.6258.
Roy, Deb K. 2002. “Learning Visually Grounded Words and Syntax for a Scene Description Task.” Computer Speech &Amp; Language 16 (July). https://doi.org/10.1016/s0885-2308(02)00024-4.
Roy, Subhankar, Aliaksandr Siarohin, Enver Sangineto, Samuel Rota Bulo, Nicu Sebe, and Elisa Ricci. 2019. “Unsupervised Domain Adaptation Using Feature-Whitening and Consensus Loss.” arXiv. https://doi.org/10.48550/ARXIV.1903.03215.
Royle, J. Andrew, Richard B. Chandler, Charles Yackulic, and James D. Nichols. 2012. “Likelihood Analysis of Species Occurrence Probability from Presence‐only Data for Modelling Species Distributions.” Methods in Ecology and Evolution 3 (January). https://doi.org/10.1111/j.2041-210x.2011.00182.x.
Rozsa, Andras, Ethan M. Rudd, and Terrance E. Boult. 2016. “Adversarial Diversity and Hard Positive Generation.” arXiv. https://doi.org/10.48550/ARXIV.1605.01775.
Rubin, Timothy N., America Chambers, Padhraic Smyth, and Mark Steyvers. 2011. “Statistical Topic Models for Multi-Label Document Classification.” arXiv. https://doi.org/10.48550/ARXIV.1107.2462.
Ruder, Sebastian. 2017. “An Overview of Multi-Task Learning in Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1706.05098.
Rue, Håvard, Sara Martino, and Nicolas Chopin. 2009. “Approximate Bayesian Inference for Latent Gaussian Models by Using Integrated Nested Laplace Approximations.” Journal of the Royal Statistical Society Series B: Statistical Methodology 71 (April). https://doi.org/10.1111/j.1467-9868.2008.00700.x.
Ruggieri, Salvatore. 2012. “Subtree Replacement in Decision Tree Simplification.” Proceedings of the 2012 SIAM International Conference on Data Mining, April. https://doi.org/10.1137/1.9781611972825.33.
Rumberg, Lars, Christopher Gebauer, Hanna Ehlert, Maren Wallbaum, Ulrike Lüdtke, and Joern Ostermann. 2023. “Uncertainty Estimation for Connectionist Temporal Classification Based Automatic Speech Recognition.” INTERSPEECH 2023, August. https://doi.org/10.21437/interspeech.2023-907.
Rumelhart, David E., Geoffrey E. Hinton, and Ronald J. Williams. 1986. “Learning Representations by Back-Propagating Errors.” Nature 323 (October). https://doi.org/10.1038/323533a0.
Rushmore, R. Jarrett, Kyle Sutherland, Holly Carrington, Justine Chen, Michael Halle, Andras Lasso, George Papadimitriou, et al. 2022. “HOA-2/SubcorticalParcellations: Release-50-Subjects-1.0.0,” August. https://doi.org/10.5281/ZENODO.6967315.
Rusk, Nicole. 2015. “Deep Learning.” Nature Methods 13 (December). https://doi.org/10.1038/nmeth.3707.
Ruskov, Martin. 2023. “Grimm in Wonderland: Prompt Engineering with Midjourney to Illustrate Fairytales.” arXiv. https://doi.org/10.48550/ARXIV.2302.08961.
Russakovsky, Olga, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, et al. 2014. “ImageNet Large Scale Visual Recognition Challenge.” arXiv. https://doi.org/10.48550/ARXIV.1409.0575.
———, et al. 2015. “ImageNet Large Scale Visual Recognition Challenge.” International Journal of Computer Vision 115 (April). https://doi.org/10.1007/s11263-015-0816-y.
Russell, Rebecca L., Louis Kim, Lei H. Hamilton, Tomo Lazovich, Jacob A. Harer, Onur Ozdemir, Paul M. Ellingwood, and Marc W. McConley. 2018. “Automated Vulnerability Detection in Source Code Using Deep Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1807.04320.
Russo, Francesco, and Pierre Vallois. 1993. “Forward, Backward and Symmetric Stochastic Integration.” Probability Theory and Related Fields 97 (September). https://doi.org/10.1007/bf01195073.
Rusu, Andrei A., Neil C. Rabinowitz, Guillaume Desjardins, Hubert Soyer, James Kirkpatrick, Koray Kavukcuoglu, Razvan Pascanu, and Raia Hadsell. 2016. “Progressive Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1606.04671.
Ruthotto, Lars, and Eldad Haber. 2018. “Deep Neural Networks Motivated by Partial Differential Equations.” arXiv. https://doi.org/10.48550/ARXIV.1804.04272.
Ryffel, Theo, Andrew Trask, Morten Dahl, Bobby Wagner, Jason Mancuso, Daniel Rueckert, and Jonathan Passerat-Palmbach. 2018. “A Generic Framework for Privacy Preserving Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1811.04017.
Rzepakowski, Piotr, and Szymon Jaroszewicz. 2011. “Decision Trees for Uplift Modeling with Single and Multiple Treatments.” Knowledge and Information Systems 32 (July). https://doi.org/10.1007/s10115-011-0434-0.
Sa, Christopher De, Albert Gu, Christopher Ré, and Frederic Sala. 2018. “Representation Tradeoffs for Hyperbolic Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1804.03329.
Sabater, Jordi, and Carles Sierra. 2005. “Review on Computational Trust and Reputation Models.” Artificial Intelligence Review 24 (September). https://doi.org/10.1007/s10462-004-0041-5.
Sachs, Matthias, Deborshee Sen, Jianfeng Lu, and David Dunson. 2023. “Posterior Computation with the Gibbs Zig-Zag Sampler.” Bayesian Analysis 18 (September). https://doi.org/10.1214/22-ba1319.
Sadasivan, Vinu Sankar, Aounon Kumar, Sriram Balasubramanian, Wenxiao Wang, and Soheil Feizi. 2023. “Can AI-Generated Text Be Reliably Detected?” arXiv. https://doi.org/10.48550/ARXIV.2303.11156.
Sadygov, Rovshan G, Daniel Cociorva, and John R Yates. 2004. “Large-Scale Database Searching Using Tandem Mass Spectra: Looking up the Answer in the Back of the Book.” Nature Methods 1 (November). https://doi.org/10.1038/nmeth725.
Saha, Rohit, Brendan Duke, Florian Shkurti, Graham W. Taylor, and Parham Aarabi. 2021. “LOHO: Latent Optimization of Hairstyles via Orthogonalization.” arXiv. https://doi.org/10.48550/ARXIV.2103.03891.
Sainath, Tara N., Chung-Cheng Chiu, Rohit Prabhavalkar, Anjuli Kannan, Yonghui Wu, Patrick Nguyen, and Zhifeng Chen. 2017. “Improving the Performance of Online Neural Transducer Models.” arXiv. https://doi.org/10.48550/ARXIV.1712.01807.
Sainath, Tara N., Rohit Prabhavalkar, Ankur Bapna, Yu Zhang, Zhouyuan Huo, Zhehuai Chen, Bo Li, Weiran Wang, and Trevor Strohman. 2022. “JOIST: A Joint Speech and Text Streaming Model for ASR.” arXiv. https://doi.org/10.48550/ARXIV.2210.07353.
Saisho, Yasumasa. 1987. “Stochastic Differential Equations for Multi-Dimensional Domain with Reflecting Boundary.” Probability Theory and Related Fields 74 (September). https://doi.org/10.1007/bf00699100.
Saito, Morihiko. 2007. “On Real Log Canonical Thresholds.” arXiv. https://doi.org/10.48550/ARXIV.0707.2308.
Sak, Haşim, Andrew Senior, and Françoise Beaufays. 2014. “Long Short-Term Memory Based Recurrent Neural Network Architectures for Large Vocabulary Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1402.1128.
Sak, Haşim, Andrew Senior, Kanishka Rao, and Françoise Beaufays. 2015. “Fast and Accurate Recurrent Neural Network Acoustic Models for Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1507.06947.
Sakkis, G., I. Androutsopoulos, G. Paliouras, V. Karkaletsis, C. D. Spyropoulos, and P. Stamatopoulos. 2001. “Stacking Classifiers for Anti-Spam Filtering of e-Mail.” arXiv. https://doi.org/10.48550/ARXIV.CS/0106040.
Sakti, Sakriani, Michael Paul, Andrew Finch, Xinhui Hu, Jinfu Ni, Noriyuki Kimura, Shigeki Matsuda, et al. 2012. “Distributed Speech Translation Technologies for Multiparty Multilingual Communication.” ACM Transactions on Speech and Language Processing 9 (July). https://doi.org/10.1145/2287710.2287712.
Salami, Behzad, Osman Unsal, and Adrian Cristal. 2018. “On the Resilience of RTL NN Accelerators: Fault Characterization and Mitigation.” arXiv. https://doi.org/10.48550/ARXIV.1806.09679.
Saleh, Hind, Areej Alhothali, and Kawthar Moria. 2023. “Detection of Hate Speech Using BERT and Hate Speech Word Embedding with Deep Model.” Applied Artificial Intelligence 37 (February). https://doi.org/10.1080/08839514.2023.2166719.
Saleh, Majd, and Stéphane Paquelet. 2024. “Anatomy of Neural Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.03797.
Salehi, Seyed Sadegh Mohseni, Deniz Erdogmus, and Ali Gholipour. 2017. “Tversky Loss Function for Image Segmentation Using 3D Fully Convolutional Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1706.05721.
Salimans, Tim, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. 2016. “Improved Techniques for Training GANs.” arXiv. https://doi.org/10.48550/ARXIV.1606.03498.
Sallam, Malik. 2023. “The Utility of ChatGPT as an Example of Large Language Models in Healthcare Education, Research and Practice: Systematic Review on the Future Perspectives and Potential Limitations,” February. https://doi.org/10.1101/2023.02.19.23286155.
Sallam, Malik, Nesreen A Salim, Ala’a B Al-Tammemi, Muna Barakat, Diaa Fayyad, Souheil Hallit, Harapan Harapan, Rabih Hallit, and Azmi Mahafzah. 2023. “ChatGPT Output Regarding Compulsory Vaccination and COVID-19 Vaccine Conspiracy: A Descriptive Study at the Outset of a Paradigm Shift in Online Search for Information.” Cureus, February. https://doi.org/10.7759/cureus.35029.
Saltori, Cristiano, Fabio Galasso, Giuseppe Fiameni, Nicu Sebe, Elisa Ricci, and Fabio Poiesi. 2022. “CoSMix: Compositional Semantic Mix for Domain Adaptation in 3D LiDAR Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.2207.09778.
Samangouei, Pouya, Maya Kabkab, and Rama Chellappa. 2018. “Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.1805.06605.
Samavatian, Mohammad Hossein, Anys Bacha, Li Zhou, and Radu Teodorescu. 2020. “RNNFast.” ACM Journal on Emerging Technologies in Computing Systems 16 (September). https://doi.org/10.1145/3399670.
Samek, Wojciech, Thomas Wiegand, and Klaus-Robert Müller. 2017. “Explainable Artificial Intelligence: Understanding, Visualizing and Interpreting Deep Learning Models.” arXiv. https://doi.org/10.48550/ARXIV.1708.08296.
Samin, Ahnaf Mozib, Behrooz Nikandish, and Jingyan Chen. 2022. “Arguments to Key Points Mapping with Prompt-Based Learning.” arXiv. https://doi.org/10.48550/ARXIV.2211.14995.
Sanchez-Vega, Francisco, Jason Eisner, Laurent Younes, and Donald Geman. 2013. “Learning Multivariate Distributions by Competitive Assembly of Marginals.” IEEE Transactions on Pattern Analysis and Machine Intelligence 35 (February). https://doi.org/10.1109/tpami.2012.96.
Sanderson, Katharine. 2023. “GPT-4 Is Here: What Scientists Think.” Nature 615 (March). https://doi.org/10.1038/d41586-023-00816-5.
Sandler, Mark, Andrey Zhmoginov, Max Vladymyrov, and Andrew Jackson. 2022. “Fine-Tuning Image Transformers Using Learnable Memory.” arXiv. https://doi.org/10.48550/ARXIV.2203.15243.
Sang, Huiyan, Mikyoung Jun, and Jianhua Z. Huang. 2011. “Covariance Approximation for Large Multivariate Spatial Data Sets with an Application to Multiple Climate Model Errors.” The Annals of Applied Statistics 5 (December). https://doi.org/10.1214/11-aoas478.
Sanh, Victor, Thomas Wolf, and Sebastian Ruder. 2018. “A Hierarchical Multi-Task Approach for Learning Embeddings from Semantic Tasks.” arXiv. https://doi.org/10.48550/ARXIV.1811.06031.
Sankar, Aravind, Yanhong Wu, Liang Gou, Wei Zhang, and Hao Yang. 2020. “DySAT.” Proceedings of the 13th International Conference on Web Search and Data Mining, January. https://doi.org/10.1145/3336191.3371845.
Santerne, A., C. Moutou, M. Tsantaki, F. Bouchy, G. Hébrard, V. Adibekyan, J.-M. Almenara, et al. 2016. “SOPHIE Velocimetry of<i>kepler</i>transit Candidates.” Astronomy &Amp; Astrophysics 587 (February). https://doi.org/10.1051/0004-6361/201527329.
Santoro, Adam, Sergey Bartunov, Matthew Botvinick, Daan Wierstra, and Timothy Lillicrap. 2016. “One-Shot Learning with Memory-Augmented Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1605.06065.
Santu, Shubhra Kanti Karmaker, and Dongji Feng. 2023. “TELeR: A General Taxonomy of LLM Prompts for Benchmarking Complex Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2305.11430.
Sariyar, Murat, and Andreas Borg. 2010. “The RecordLinkage Package: Detecting Errors in Data.” The R Journal 2. https://doi.org/10.32614/rj-2010-017.
Särkkä, Simo. 2013. “Bayesian Filtering and Smoothing,” September. https://doi.org/10.1017/cbo9781139344203.
Sato, Ken-iti. 1991. “Self-Similar Processes with Independent Increments.” Probability Theory and Related Fields 89 (September). https://doi.org/10.1007/bf01198788.
Sattler, Felix, Klaus-Robert Müller, and Wojciech Samek. 2019. “Clustered Federated Learning: Model-Agnostic Distributed Multi-Task Optimization Under Privacy Constraints.” arXiv. https://doi.org/10.48550/ARXIV.1910.01991.
Sattler, Felix, Simon Wiedemann, Klaus-Robert Müller, and Wojciech Samek. 2019. “Robust and Communication-Efficient Federated Learning from Non-IID Data.” arXiv. https://doi.org/10.48550/ARXIV.1903.02891.
Sau, Bharat Bhusan, and Vineeth N. Balasubramanian. 2016. “Deep Model Compression: Distilling Knowledge from Noisy Teachers.” arXiv. https://doi.org/10.48550/ARXIV.1610.09650.
Sauer, Axel, Tero Karras, Samuli Laine, Andreas Geiger, and Timo Aila. 2023. “StyleGAN-t: Unlocking the Power of GANs for Fast Large-Scale Text-to-Image Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.2301.09515.
Savage, David, Qingmai Wang, Pauline Chou, Xiuzhen Zhang, and Xinghuo Yu. 2016. “Detection of Money Laundering Groups Using Supervised Learning in Networks.” arXiv. https://doi.org/10.48550/ARXIV.1608.00708.
Sawhney, Ramit, Megh Thakkar, Shivam Agarwal, Di Jin, Diyi Yang, and Lucie Flek. 2021. “HypMix: Hyperbolic Interpolative Data Augmentation.” Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/2021.emnlp-main.776.
Sayed, Ali. 2014. “Adaptation, Learning, and Optimization over Networks.” Foundations and Trends® in Machine Learning 7. https://doi.org/10.1561/2200000051.
“Scalable Optimization via Probabilistic Modeling.” 2006. Studies in Computational Intelligence. https://doi.org/10.1007/978-3-540-34954-9.
Scarlini, Bianca, Tommaso Pasini, and Roberto Navigli. 2019. “Just ‘OneSeC’ for Producing Multilingual Sense-Annotated Data.” Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/p19-1069.
Schaeffer, Rylan, Brando Miranda, and Sanmi Koyejo. 2023. “Are Emergent Abilities of Large Language Models a Mirage?” arXiv. https://doi.org/10.48550/ARXIV.2304.15004.
Scheurer, Jérémy, Jon Ander Campos, Tomasz Korbak, Jun Shern Chan, Angelica Chen, Kyunghyun Cho, and Ethan Perez. 2023. “Training Language Models with Language Feedback at Scale.” arXiv. https://doi.org/10.48550/ARXIV.2303.16755.
Schick, Timo, Jane Dwivedi-Yu, Roberto Dessì, Roberta Raileanu, Maria Lomeli, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. 2023. “Toolformer: Language Models Can Teach Themselves to Use Tools.” arXiv. https://doi.org/10.48550/ARXIV.2302.04761.
Schindler, Guenther, Wolfgang Roth, Franz Pernkopf, and Holger Froening. 2019. “Parameterized Structured Pruning for Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1906.05180.
Schmaltz, Allen. 2019. “Detecting Local Insights from Global Labels: Supervised &Amp; Zero-Shot Sequence Labeling via a Convolutional Decomposition.” arXiv. https://doi.org/10.48550/ARXIV.1906.01154.
Schmidt, Philip, Attila Reiss, Robert Duerichen, Claus Marberger, and Kristof Van Laerhoven. 2018. “Introducing WESAD, a Multimodal Dataset for Wearable Stress and Affect Detection.” Proceedings of the 20th ACM International Conference on Multimodal Interaction, October. https://doi.org/10.1145/3242969.3242985.
Schmidt, Robin M., Frank Schneider, and Philipp Hennig. 2020. “Descending Through a Crowded Valley - Benchmarking Deep Learning Optimizers.” arXiv. https://doi.org/10.48550/ARXIV.2007.01547.
Schneider, Johannes, Steffi Haag, and Leona Chandra Kruse. 2023. “Negotiating with LLMS: Prompt Hacks, Skill Gaps, and Reasoning Deficits.” arXiv. https://doi.org/10.48550/ARXIV.2312.03720.
Schölkopf, Bernhard, John C. Platt, John Shawe-Taylor, Alex J. Smola, and Robert C. Williamson. 2001. “Estimating the Support of a High-Dimensional Distribution.” Neural Computation 13 (July). https://doi.org/10.1162/089976601750264965.
Schulhoff, Sander, Jeremy Pinto, Anaum Khan, Louis-François Bouchard, Chenglei Si, Svetlina Anati, Valen Tagliabue, Anson Liu Kost, Christopher Carnahan, and Jordan Boyd-Graber. 2023. “Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of LLMs Through a Global Scale Prompt Hacking Competition.” arXiv. https://doi.org/10.48550/ARXIV.2311.16119.
Schulman, John, Xi Chen, and Pieter Abbeel. 2017. “Equivalence Between Policy Gradients and Soft q-Learning.” arXiv. https://doi.org/10.48550/ARXIV.1704.06440.
Schulman, John, Sergey Levine, Philipp Moritz, Michael I. Jordan, and Pieter Abbeel. 2015. “Trust Region Policy Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1502.05477.
Schuurmans, Dale. 2023. “Memory Augmented Large Language Models Are Computationally Universal.” arXiv. https://doi.org/10.48550/ARXIV.2301.04589.
Schwaller, Philippe, Théophile Gaudin, Dávid Lányi, Costas Bekas, and Teodoro Laino. 2018. ‘Found in Translation’: Predicting Outcomes of Complex Organic Chemistry Reactions Using Neural Sequence-to-Sequence Models.” Chemical Science 9. https://doi.org/10.1039/c8sc02339e.
Schwartz, Howard M. 2014. “Multi‐agent Machine Learning,” May. https://doi.org/10.1002/9781118884614.
Schwartz, Michael P., Zhonggang Hou, Nicholas E. Propson, Jue Zhang, Collin J. Engstrom, Vitor Santos Costa, Peng Jiang, et al. 2015. “Human Pluripotent Stem Cell-Derived Neural Constructs for Predicting Neural Toxicity.” Proceedings of the National Academy of Sciences 112 (September). https://doi.org/10.1073/pnas.1516645112.
Scutari, Marco. 2009. “Learning Bayesian Networks with the Bnlearn r Package.” arXiv. https://doi.org/10.48550/ARXIV.0908.3817.
Sedhain, Suvash, Aditya Krishna Menon, Scott Sanner, and Lexing Xie. 2015. “AutoRec.” Proceedings of the 24th International Conference on World Wide Web, May. https://doi.org/10.1145/2740908.2742726.
Sehwag, Vikash, Saeed Mahloujifar, Tinashe Handina, Sihui Dai, Chong Xiang, Mung Chiang, and Prateek Mittal. 2021. “Robust Learning Meets Generative Models: Can Proxy Distributions Improve Adversarial Robustness?” arXiv. https://doi.org/10.48550/ARXIV.2104.09425.
Sejdinovic, Dino, Bharath Sriperumbudur, Arthur Gretton, and Kenji Fukumizu. 2013. “Equivalence of Distance-Based and RKHS-Based Statistics in Hypothesis Testing.” The Annals of Statistics 41 (October). https://doi.org/10.1214/13-aos1140.
Sejnowski, T. J. 1977. “Storing Covariance with Nonlinearly Interacting Neurons.” Journal of Mathematical Biology 4. https://doi.org/10.1007/bf00275079.
“Semi-Supervised Learning.” 2006, September. https://doi.org/10.7551/mitpress/9780262033589.001.0001.
Senadeera, Damith Chamalke, and Julia Ive. 2022. “Controlled Text Generation Using T5 Based Encoder-Decoder Soft Prompt Tuning and Analysis of the Utility of Generated Text in AI.” arXiv. https://doi.org/10.48550/ARXIV.2212.02924.
Senekane, Makhamisa, and Benedict Molibeli Taele. 2016. “Prediction of Solar Irradiation Using Quantum Support Vector Machine Learning Algorithm.” Smart Grid and Renewable Energy 07. https://doi.org/10.4236/sgre.2016.712022.
Sengupta, Abhronil, Yuting Ye, Robert Wang, Chiao Liu, and Kaushik Roy. 2018. “Going Deeper in Spiking Neural Networks: VGG and Residual Architectures.” arXiv. https://doi.org/10.48550/ARXIV.1802.02627.
Seo, Minjoon, Aniruddha Kembhavi, Ali Farhadi, and Hannaneh Hajishirzi. 2016. “Bidirectional Attention Flow for Machine Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.1611.01603.
Seo, Minjoon, Sewon Min, Ali Farhadi, and Hannaneh Hajishirzi. 2016. “Query-Reduction Networks for Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1606.04582.
Serban, Iulian Vlad, Alberto García-Durán, Caglar Gulcehre, Sungjin Ahn, Sarath Chandar, Aaron Courville, and Yoshua Bengio. 2016. “Generating Factoid Questions with Recurrent Neural Networks: The 30M Factoid Question-Answer Corpus.” arXiv. https://doi.org/10.48550/ARXIV.1603.06807.
Serban, Iulian Vlad, Tim Klinger, Gerald Tesauro, Kartik Talamadupula, Bowen Zhou, Yoshua Bengio, and Aaron Courville. 2016. “Multiresolution Recurrent Neural Networks: An Application to Dialogue Response Generation.” arXiv. https://doi.org/10.48550/ARXIV.1606.00776.
Serban, Iulian Vlad, Ryan Lowe, Laurent Charlin, and Joelle Pineau. 2016. “Generative Deep Neural Networks for Dialogue: A Short Review.” arXiv. https://doi.org/10.48550/ARXIV.1611.06216.
Sermanet, Pierre, Soumith Chintala, and Yann LeCun. 2012. “Convolutional Neural Networks Applied to House Numbers Digit Classification.” arXiv. https://doi.org/10.48550/ARXIV.1204.3968.
Sermanet, Pierre, David Eigen, Xiang Zhang, Michael Mathieu, Rob Fergus, and Yann LeCun. 2013. “OverFeat: Integrated Recognition, Localization and Detection Using Convolutional Networks,” December. http://arxiv.org/abs/1312.6229v4.
Seth, Anil K. 2015. “The Cybernetic Bayesian Brain.” Open MIND. https://doi.org/10.15502/9783958570108.
Sezener, Eren, Agnieszka Grabska-Barwińska, Dimitar Kostadinov, Maxime Beau, Sanjukta Krishnagopal, David Budden, Marcus Hutter, et al. 2021. “A Rapid and Efficient Learning Rule for Biological Neural Circuits,” March. https://doi.org/10.1101/2021.03.10.434756.
Sgarbossa, Damiano, Umberto Lupo, and Anne-Florence Bitbol. 2023. “Generative Power of a Protein Language Model Trained on Multiple Sequence Alignments.” eLife 12 (February). https://doi.org/10.7554/elife.79854.
Shabalin, Andrey A., Victor J. Weigman, Charles M. Perou, and Andrew B. Nobel. 2009. “Finding Large Average Submatrices in High Dimensional Data.” The Annals of Applied Statistics 3 (September). https://doi.org/10.1214/09-aoas239.
Shah, Parantu K, Carolina Perez-Iratxeta, Peer Bork, and Miguel A Andrade. 2003. “Information Extraction from Full Text Scientific Articles: Where Are the Keywords?” BMC Bioinformatics 4 (May). https://doi.org/10.1186/1471-2105-4-20.
Shaham, Uri, Elad Segal, Maor Ivgi, Avia Efrat, Ori Yoran, Adi Haviv, Ankit Gupta, et al. 2022. “SCROLLS: Standardized CompaRison over Long Language Sequences.” arXiv. https://doi.org/10.48550/ARXIV.2201.03533.
Shalev-Shwartz, Shai, Shaked Shammah, and Amnon Shashua. 2016. “Safe, Multi-Agent, Reinforcement Learning for Autonomous Driving.” arXiv. https://doi.org/10.48550/ARXIV.1610.03295.
Shalev-Shwartz, Shai, and Tong Zhang. 2013. “Accelerated Proximal Stochastic Dual Coordinate Ascent for Regularized Loss Minimization.” arXiv. https://doi.org/10.48550/ARXIV.1309.2375.
Shamir, Ohad, and Tong Zhang. 2012. “Stochastic Gradient Descent for Non-Smooth Optimization: Convergence Results and Optimal Averaging Schemes.” arXiv. https://doi.org/10.48550/ARXIV.1212.1824.
Shan, Lili, Lei Lin, Chengjie Sun, and Xiaolong Wang. 2016. “Predicting Ad Click-Through Rates via Feature-Based Fully Coupled Interaction Tensor Factorization.” Electronic Commerce Research and Applications 16 (March). https://doi.org/10.1016/j.elerap.2016.01.004.
Shang, Junyuan, Tengfei Ma, Cao Xiao, and Jimeng Sun. 2019. “Pre-Training of Graph Augmented Transformers for Medication Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.1906.00346.
Shannon, Matt. 2017. “Optimizing Expected Word Error Rate via Sampling for Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1706.02776.
Shao, Wujun, Pengli Ji, Dongwei Fan, Yaohua Hu, Xiaoran Yan, Chenzhou Cui, Linying Mi, Lang Chen, and Rui Zhang. 2023. “Astronomical Knowledge Entity Extraction in Astrophysics Journal Articles via Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2310.17892.
Shao, Xuhui, and Lexin Li. 2011. “Data-Driven Multi-Touch Attribution Models.” Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2020408.2020453.
Shapovalova, Yuliya, Tom Heskes, and Tjeerd Dijkstra. 2022. “Non-Parametric Synergy Modeling of Chemical Compounds with Gaussian Processes.” BMC Bioinformatics 23 (January). https://doi.org/10.1186/s12859-021-04508-7.
Sharma, Ashish, Inna W. Lin, Adam S. Miner, David C. Atkins, and Tim Althoff. 2021. “Towards Facilitating Empathic Conversations in Online Mental Health Support: A Reinforcement Learning Approach.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3450097.
Shatkay, Hagit, Annette Höglund, Scott Brady, Torsten Blum, Pierre Dönnes, and Oliver Kohlbacher. 2007. “SherLoc: High-Accuracy Prediction of Protein Subcellular Localization by Integrating Text and Protein Sequence Data.” Bioinformatics 23 (March). https://doi.org/10.1093/bioinformatics/btm115.
Shavit, Yonadav. 2023. “What Does It Take to Catch a Chinchilla? Verifying Rules on Large-Scale Neural Network Training via Compute Monitoring.” arXiv. https://doi.org/10.48550/ARXIV.2303.11341.
Shawahna, Ahmad, Sadiq M. Sait, and Aiman El-Maleh. 2019. “FPGA-Based Accelerators of Deep Learning Networks for Learning and Classification: A Review.” IEEE Access 7. https://doi.org/10.1109/access.2018.2890150.
Shawe-Taylor, John, and Robert C. Williamson. 1997. “A PAC Analysis of a Bayesian Estimator.” Proceedings of the Tenth Annual Conference on Computational Learning Theory - COLT ’97. https://doi.org/10.1145/267460.267466.
Sheen, Spencer, and Jiancheng Lyu. 2019. “Median Binary-Connect Method and a Binary Convolutional Neural Network for Word Recognition.” 2019 Data Compression Conference (DCC), March. https://doi.org/10.1109/dcc.2019.00116.
Shehnepoor, Saeedreza, Roberto Togneri, Wei Liu, and Mohammed Bennamoun. 2020. “DFraud3- Multi-Component Fraud Detection Freeof Cold-Start.” arXiv. https://doi.org/10.48550/ARXIV.2006.05718.
Sheikh, Nasrullah, Xiao Qin, Yaniv Gur, and Berthold Reinwald. 2022. “Distributed Training of Knowledge Graph Embedding Models Using Ray.” https://doi.org/10.48786/EDBT.2022.48.
Sheikh, Shakeel Ahmad, Md Sahidullah, Fabrice Hirsch, and Slim Ouni. 2022. “Introducing ECAPA-TDNN and Wav2Vec2.0 Embeddings to Stuttering Detection.” arXiv. https://doi.org/10.48550/ARXIV.2204.01564.
Shekhar, Shubhranshu, Neil Shah, and Leman Akoglu. 2021. “FairOD: Fairness-Aware Outlier Detection.” Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, July. https://doi.org/10.1145/3461702.3462517.
Shelah, Saharon. 1974. “Infinite Abelian Groups, Whitehead Problem and Some Constructions.” Israel Journal of Mathematics 18 (September). https://doi.org/10.1007/bf02757281.
Shen, Aili, Xudong Han, Trevor Cohn, Timothy Baldwin, and Lea Frermann. 2022. “Optimising Equal Opportunity Fairness in Model Training.” arXiv. https://doi.org/10.48550/ARXIV.2205.02393.
Shen, Jonathan, Noranart Vesdapunt, Vishnu N. Boddeti, and Kris M. Kitani. 2016. “In Teacher We Trust: Learning Compressed Models for Pedestrian Detection.” arXiv. https://doi.org/10.48550/ARXIV.1612.00478.
Shen, Sheng, Liunian Harold Li, Hao Tan, Mohit Bansal, Anna Rohrbach, Kai-Wei Chang, Zhewei Yao, and Kurt Keutzer. 2021. “How Much Can CLIP Benefit Vision-and-Language Tasks?” arXiv. https://doi.org/10.48550/ARXIV.2107.06383.
Shen, Wen, Binbin Zhang, Shikun Huang, Zhihua Wei, and Quanshi Zhang. 2019. “3D-Rotation-Equivariant Quaternion Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1911.09040.
Shen, Xinyue, Zeyuan Chen, Michael Backes, and Yang Zhang. 2023. “In ChatGPT We Trust? Measuring and Characterizing the Reliability of ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2304.08979.
Shen, Xuan, Yaohua Wang, Ming Lin, Yilun Huang, Hao Tang, Xiuyu Sun, and Yanzhi Wang. 2023. “DeepMAD: Mathematical Architecture Design for Deep Convolutional Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.2303.02165.
Shen, Yang, Xuhao Sun, and Xiu-Shen Wei. 2023. “Equiangular Basis Vectors.” arXiv. https://doi.org/10.48550/ARXIV.2303.11637.
Shen, Yifei, Jiawei Shao, Xinjie Zhang, Zehong Lin, Hao Pan, Dongsheng Li, Jun Zhang, and Khaled B. Letaief. 2023. “Large Language Models Empowered Autonomous Edge AI for Connected Intelligence.” arXiv. https://doi.org/10.48550/ARXIV.2307.02779.
Shen, Yongliang, Kaitao Song, Xu Tan, Dongsheng Li, Weiming Lu, and Yueting Zhuang. 2023. “HuggingGPT: Solving AI Tasks with ChatGPT and Its Friends in Hugging Face.” arXiv. https://doi.org/10.48550/ARXIV.2303.17580.
Shen, Zejiang, Ruochen Zhang, Melissa Dell, Benjamin Charles Germain Lee, Jacob Carlson, and Weining Li. 2021. “LayoutParser: A Unified Toolkit for Deep Learning Based Document Image Analysis.” arXiv. https://doi.org/10.48550/ARXIV.2103.15348.
Sheng, Ying, Lianmin Zheng, Binhang Yuan, Zhuohan Li, Max Ryabinin, Daniel Y. Fu, Zhiqiang Xie, et al. 2023. “FlexGen: High-Throughput Generative Inference of Large Language Models with a Single GPU.” arXiv. https://doi.org/10.48550/ARXIV.2303.06865.
Sherkat, Ehsan, and Evangelos Milios. 2017. “Vector Embedding of Wikipedia Concepts and Entities.” arXiv. https://doi.org/10.48550/ARXIV.1702.03470.
Sheth, Amit, Kaushik Roy, and Manas Gaur. 2023. “Neurosymbolic AI – Why, What, and How.” arXiv. https://doi.org/10.48550/ARXIV.2305.00813.
Sheynin, Shelly, Oron Ashual, Adam Polyak, Uriel Singer, Oran Gafni, Eliya Nachmani, and Yaniv Taigman. 2022. “KNN-Diffusion: Image Generation via Large-Scale Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.2204.02849.
Shi, Bowen, Ming Sun, Chieh-Chi Kao, Viktor Rozgic, Spyros Matsoukas, and Chao Wang. 2019. “Compression of Acoustic Event Detection Models with Quantized Distillation.” arXiv. https://doi.org/10.48550/ARXIV.1907.00873.
Shi, Fengzhao, Yanan Cao, Yanmin Shang, Yuchen Zhou, Chuan Zhou, and Jia Wu. 2022. “H2-FDetector: A GNN-Based Fraud Detector with Homophilic and Heterophilic Connections.” Proceedings of the ACM Web Conference 2022, April. https://doi.org/10.1145/3485447.3512195.
Shi, Hangjie, Leslie Ball, Govind Thattai, Desheng Zhang, Lucy Hu, Qiaozi Gao, Suhaila Shakiah, et al. 2023. “Alexa, Play with Robot: Introducing the First Alexa Prize SimBot Challenge on Embodied AI.” arXiv. https://doi.org/10.48550/ARXIV.2308.05221.
Shi, Peng, and Jimmy Lin. 2019. “Simple BERT Models for Relation Extraction and Semantic Role Labeling.” arXiv. https://doi.org/10.48550/ARXIV.1904.05255.
Shi, Shaohuai, Qiang Wang, Pengfei Xu, and Xiaowen Chu. 2016. “Benchmarking State-of-the-Art Deep Learning Software Tools.” arXiv. https://doi.org/10.48550/ARXIV.1608.07249.
Shi, Weijia, Xiaochuang Han, Hila Gonen, Ari Holtzman, Yulia Tsvetkov, and Luke Zettlemoyer. 2022. “Toward Human Readable Prompt Tuning: Kubrick’s the Shining Is a Good Movie, and a Good Prompt Too?” arXiv. https://doi.org/10.48550/ARXIV.2212.10539.
Shi, Wenxuan, Fei Li, Jingye Li, Hao Fei, and Donghong Ji. 2022. “Effective Token Graph Modeling Using a Novel Labeling Strategy for Structured Sentiment Analysis.” arXiv. https://doi.org/10.48550/ARXIV.2203.10796.
Shi, Xingjian, Zhourong Chen, Hao Wang, Dit-Yan Yeung, Wai-kin Wong, and Wang-chun Woo. 2015. “Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting.” arXiv. https://doi.org/10.48550/ARXIV.1506.04214.
Shi, Yangyang, Yongqiang Wang, Chunyang Wu, Ching-Feng Yeh, Julian Chan, Frank Zhang, Duc Le, and Mike Seltzer. 2020. “Emformer: Efficient Memory Transformer Based Acoustic Model for Low Latency Streaming Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2010.10759.
Shi, Yu, Qi Zhu, Fang Guo, Chao Zhang, and Jiawei Han. 2018. “Easing Embedding Learning by Comprehensive Transcription of Heterogeneous Information Networks.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220006.
Shi, Zhengxiang, and Aldo Lipani. 2023. “DePT: Decomposed Prompt Tuning for Parameter-Efficient Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2309.05173.
Shiguihara, Pedro, Alneu De Andrade Lopes, and David Mauricio. 2021. “Dynamic Bayesian Network Modeling, Learning, and Inference: A Survey.” IEEE Access 9. https://doi.org/10.1109/access.2021.3105520.
Shih, Yi-Jen, Hsuan-Fu Wang, Heng-Jui Chang, Layne Berry, Hung-yi Lee, and David Harwath. 2022. “SpeechCLIP: Integrating Speech with Pre-Trained Vision and Language Model.” arXiv. https://doi.org/10.48550/ARXIV.2210.00705.
Shimabukuro, David W, Christopher W Barton, Mitchell D Feldman, Samson J Mataraso, and Ritankar Das. 2017. “Effect of a Machine Learning-Based Severe Sepsis Prediction Algorithm on Patient Survival and Hospital Length of Stay: A Randomised Clinical Trial.” BMJ Open Respiratory Research 4 (November). https://doi.org/10.1136/bmjresp-2017-000234.
Shin, Bonggun, Hao Yang, and Jinho D. Choi. 2019. “The Pupil Has Become the Master: Teacher-Student Model-Based Word Embedding Distillation with Ensemble Learning.” arXiv. https://doi.org/10.48550/ARXIV.1906.00095.
Shin, Jiho, Clark Tang, Tahmineh Mohati, Maleknaz Nayebi, Song Wang, and Hadi Hemmati. 2023. “Prompt Engineering or Fine Tuning: An Empirical Assessment of Large Language Models in Automated Software Engineering Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2310.10508.
Shin, Su-Jin, Kyungwoo Song, and Il-Chul Moon. 2019. “Hierarchically Clustered Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1901.09906.
Shin, Sungho, Yoonho Boo, and Wonyong Sung. 2017. “Fixed-Point Optimization of Deep Neural Networks with Adaptive Step Size Retraining.” arXiv. https://doi.org/10.48550/ARXIV.1702.08171.
———. 2019. “Knowledge Distillation for Optimization of Quantized Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1909.01688.
Shinozaki, Takashi. 2018. “Competitive Learning Enriches Learning Representation and Accelerates the Fine-Tuning of CNNs.” arXiv. https://doi.org/10.48550/ARXIV.1804.09859.
Shlens, Jonathon. 2014. “A Tutorial on Principal Component Analysis.” arXiv. https://doi.org/10.48550/ARXIV.1404.1100.
Shoeybi, Mohammad, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper, and Bryan Catanzaro. 2019. “Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism.” arXiv. https://doi.org/10.48550/ARXIV.1909.08053.
Shokri, Reza, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. 2016. “Membership Inference Attacks Against Machine Learning Models.” arXiv. https://doi.org/10.48550/ARXIV.1610.05820.
Shor, Joel, Aren Jansen, Ronnie Maor, Oran Lang, Omry Tuval, Félix de Chaumont Quitry, Marco Tagliasacchi, Ira Shavitt, Dotan Emanuel, and Yinnon Haviv. 2020. “Towards Learning a Universal Non-Semantic Representation of Speech.” Interspeech 2020, October. https://doi.org/10.21437/interspeech.2020-1242.
Shrikumar, Avanti, Peyton Greenside, Anna Shcherbina, and Anshul Kundaje. 2016. “Not Just a Black Box: Learning Important Features Through Propagating Activation Differences.” arXiv. https://doi.org/10.48550/ARXIV.1605.01713.
Shrivastava, Anshumali, and Ping Li. 2014. “Asymmetric LSH (ALSH) for Sublinear Time Maximum Inner Product Search (MIPS).” arXiv. https://doi.org/10.48550/ARXIV.1405.5869.
Shu, Dong Wook, Sung Woo Park, and Junseok Kwon. 2019. “3D Point Cloud Generative Adversarial Network Based on Tree Structured Graph Convolutions.” arXiv. https://doi.org/10.48550/ARXIV.1905.06292.
Shulman, David. 2023. “Optimization Methods in Deep Learning: A Comprehensive Overview.” arXiv. https://doi.org/10.48550/ARXIV.2302.09566.
Shum, KaShun, Shizhe Diao, and Tong Zhang. 2023. “Automatic Prompt Augmentation and Selection with Chain-of-Thought from Labeled Data.” arXiv. https://doi.org/10.48550/ARXIV.2302.12822.
Shuster, Kurt, Jack Urbanek, Emily Dinan, Arthur Szlam, and Jason Weston. 2020. “Deploying Lifelong Open-Domain Dialogue Learning.” arXiv. https://doi.org/10.48550/ARXIV.2008.08076.
Shwartz-Ziv, Ravid, and Amitai Armon. 2021. “Tabular Data: Deep Learning Is Not All You Need.” arXiv. https://doi.org/10.48550/ARXIV.2106.03253.
Shypula, Alexander, Aman Madaan, Yimeng Zeng, Uri Alon, Jacob Gardner, Milad Hashemi, Graham Neubig, Parthasarathy Ranganathan, Osbert Bastani, and Amir Yazdanbakhsh. 2023. “Learning Performance-Improving Code Edits.” arXiv. https://doi.org/10.48550/ARXIV.2302.07867.
Siam, Mennatullah, Boris Oreshkin, and Martin Jagersand. 2019. “Adaptive Masked Proxies for Few-Shot Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1902.11123.
Siddiqui, Md Amran, Alan Fern, Thomas G. Dietterich, Ryan Wright, Alec Theriault, and David W. Archer. 2018. “Feedback-Guided Anomaly Discovery via Online Optimization.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220083.
Silvestri, Stefano, Shareeful Islam, Spyridon Papastergiou, Christos Tzagkarakis, and Mario Ciampi. 2023. “A Machine Learning Approach for the NLP-Based Analysis of Cyber Threats and Vulnerabilities of the Healthcare Ecosystem.” Sensors 23 (January). https://doi.org/10.3390/s23020651.
Simmons, Mark P., Kurt M. Pickett, and Masaki Miya. 2004. “How Meaningful Are Bayesian Support Values?” Molecular Biology and Evolution 21 (January). https://doi.org/10.1093/molbev/msh014.
Simonovsky, Martin, and Nikos Komodakis. 2017. “Dynamic Edge-Conditioned Filters in Convolutional Neural Networks on Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1704.02901.
———. 2018. “GraphVAE: Towards Generation of Small Graphs Using Variational Autoencoders.” arXiv. https://doi.org/10.48550/ARXIV.1802.03480.
Simonyan, Karen, Andrea Vedaldi, and Andrew Zisserman. 2013. “Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps.” arXiv. https://doi.org/10.48550/ARXIV.1312.6034.
Simonyan, Karen, and Andrew Zisserman. 2014. “Very Deep Convolutional Networks for Large-Scale Image Recognition,” September. http://arxiv.org/abs/1409.1556v6.
Singer, Philipp, Denis Helic, Andreas Hotho, and Markus Strohmaier. 2015. “HypTrails.” Proceedings of the 24th International Conference on World Wide Web, May. https://doi.org/10.1145/2736277.2741080.
Singer, Uriel, Ido Guy, and Kira Radinsky. 2019. “Node Embedding over Temporal Graphs.” Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, August. https://doi.org/10.24963/ijcai.2019/640.
Singh, Ishika, Valts Blukis, Arsalan Mousavian, Ankit Goyal, Danfei Xu, Jonathan Tremblay, Dieter Fox, Jesse Thomason, and Animesh Garg. 2022. “ProgPrompt: Generating Situated Robot Task Plans Using Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2209.11302.
Singh, Moninder, and Marco Valtorta. 2013. “An Algorithm for the Construction of Bayesian Network Structures from Data.” arXiv. https://doi.org/10.48550/ARXIV.1303.1485.
Singh, Rahul, Maneesh Sahani, and Arthur Gretton. 2019. “Kernel Instrumental Variable Regression.” arXiv. https://doi.org/10.48550/ARXIV.1906.00232.
Singla, Deepak, Soham Chatterjee, Lavanya Ramapantulu, Andres Ussa, Bharath Ramesh, and Arindam Basu. 2020. “HyNNA: Improved Performance for Neuromorphic Vision Sensor Based Surveillance Using Hybrid Neural Network Architecture.” arXiv. https://doi.org/10.48550/ARXIV.2003.08603.
Sinha, Aman, Hongseok Namkoong, Riccardo Volpi, and John Duchi. 2017. “Certifying Some Distributional Robustness with Principled Adversarial Training.” arXiv. https://doi.org/10.48550/ARXIV.1710.10571.
Sinha, Atul Kumar, Daniele Paliotta, Bálint Máté, Sebastian Pina-Otey, John A. Raine, Tobias Golling, and François Fleuret. 2022. “SUPA: A Lightweight Diagnostic Simulator for Machine Learning in Particle Physics.” arXiv. https://doi.org/10.48550/ARXIV.2202.05012.
Sisaengsuwanchai, Khanin, Navapat Nananukul, and Mayank Kejriwal. 2023. “How Does Prompt Engineering Affect ChatGPT Performance on Unsupervised Entity Resolution?” arXiv. https://doi.org/10.48550/ARXIV.2310.06174.
sivaraman, ganesh, Nicholas Jackson, Benjamin Sanchez-Lengeling, Alvaro Vazquez-Mayagoitia, Alan Aspuru-Guzik, Venkatram Vishwanath, and Juan de Pablo. 2019. “A Diversified Machine Learning Strategy for Predicting and Understanding Molecular Melting Points,” September. https://doi.org/10.26434/chemrxiv.9914378.v1.
Skerry-Ryan, RJ, Eric Battenberg, Ying Xiao, Yuxuan Wang, Daisy Stanton, Joel Shor, Ron J. Weiss, Rob Clark, and Rif A. Saurous. 2018. “Towards End-to-End Prosody Transfer for Expressive Speech Synthesis with Tacotron.” arXiv. https://doi.org/10.48550/ARXIV.1803.09047.
Sklar, Aaron E., and Nadine B. Sarter. 1999. “Good Vibrations: Tactile Feedback in Support of Attention Allocation and Human-Automation Coordination in Event-Driven Domains.” Human Factors: The Journal of the Human Factors and Ergonomics Society 41 (December). https://doi.org/10.1518/001872099779656716.
Skosnik, P. D., F. Mirza, D. R. Gitelman, T. B. Parrish, M-M. Mesulam, and P. J. Reber. 2002. “Neural Correlates of Artificial Grammar Learning.” NeuroImage 17 (November). https://doi.org/10.1006/nimg.2002.1291.
Smagulova, Kamilya, Olga Krestinskaya, and Alex Pappachen James. 2018. “A Memristor-Based Long Short Term Memory Circuit.” Analog Integrated Circuits and Signal Processing 95 (April). https://doi.org/10.1007/s10470-018-1180-y.
Smeaton, Alan F., and Jamie Callan. 2005. “Personalisation and Recommender Systems in Digital Libraries.” International Journal on Digital Libraries 5 (August). https://doi.org/10.1007/s00799-004-0100-1.
Smith, Kandler, and Paul Gasper. 2021. “AI-Batt (Autonomous Identification of Battery Life Models) [SWR 21-36].” https://doi.org/10.11578/DC.20220415.1.
Snell, Jake, Kevin Swersky, and Richard S. Zemel. 2017. “Prototypical Networks for Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.1703.05175.
Snoek, Jasper, Oren Rippel, Kevin Swersky, Ryan Kiros, Nadathur Satish, Narayanan Sundaram, Md. Mostofa Ali Patwary, and Ryan P. Adams. 2015. “Scalable Bayesian Optimization Using Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1502.05700.
So, David R., Chen Liang, and Quoc V. Le. 2019. “The Evolved Transformer.” arXiv. https://doi.org/10.48550/ARXIV.1901.11117.
Sobania, Dominik, Martin Briesch, Carol Hanna, and Justyna Petke. 2023. “An Analysis of the Automatic Bug Fixing Performance of ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2301.08653.
Sohl-Dickstein, Jascha, and Kenji Kawaguchi. 2019. “Eliminating All Bad Local Minima from Loss Landscapes Without Even Adding an Extra Unit.” arXiv. https://doi.org/10.48550/ARXIV.1901.03909.
Solaiman, Irene. 2023. “The Gradient of Generative AI Release: Methods and Considerations.” arXiv. https://doi.org/10.48550/ARXIV.2302.04844.
Solan, Zach, David Horn, Eytan Ruppin, and Shimon Edelman. 2005. “Unsupervised Learning of Natural Languages.” Proceedings of the National Academy of Sciences 102 (August). https://doi.org/10.1073/pnas.0409746102.
Soldani, Jacopo, and Antonio Brogi. 2021. “Anomaly Detection and Failure Root Cause Analysis in (Micro)service-Based Cloud Applications: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2105.12378.
Sønderby, Casper Kaae, Lasse Espeholt, Jonathan Heek, Mostafa Dehghani, Avital Oliver, Tim Salimans, Shreya Agrawal, Jason Hickey, and Nal Kalchbrenner. 2020. “MetNet: A Neural Weather Model for Precipitation Forecasting.” arXiv. https://doi.org/10.48550/ARXIV.2003.12140.
Soner, H. Mete, Nizar Touzi, and Jianfeng Zhang. 2011. “Wellposedness of Second Order Backward SDEs.” Probability Theory and Related Fields 153 (February). https://doi.org/10.1007/s00440-011-0342-y.
Song, Chuanbiao, Kun He, Liwei Wang, and John E. Hopcroft. 2018. “Improving the Generalization of Adversarial Training with Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1810.00740.
Song, Jaewoo, and Fangzhen Lin. 2022. “PocketNN: Integer-Only Training and Inference of Neural Networks via Direct Feedback Alignment and Pocket Activations in Pure c++.” arXiv. https://doi.org/10.48550/ARXIV.2201.02863.
Song, Kaitao, Xu Tan, Tao Qin, Jianfeng Lu, and Tie-Yan Liu. 2019. “MASS: Masked Sequence to Sequence Pre-Training for Language Generation.” arXiv. https://doi.org/10.48550/ARXIV.1905.02450.
Song, Le, Alex Smola, Arthur Gretton, Karsten Borgwardt, and Justin Bedo. 2007. “Supervised Feature Selection via Dependence Estimation.” arXiv. https://doi.org/10.48550/ARXIV.0704.2668.
Song, Linfeng, and Lin Zhao. 2016. “Question Generation from a Knowledge Base with Web Exploration.” arXiv. https://doi.org/10.48550/ARXIV.1610.03807.
Song, Linxin, Jieyu Zhang, Lechao Cheng, Pengyuan Zhou, Tianyi Zhou, and Irene Li. 2023. “NLPBench: Evaluating Large Language Models on Solving NLP Problems.” arXiv. https://doi.org/10.48550/ARXIV.2309.15630.
Song, Xiangxiang, Guang Ling, Wenhui Tu, and Yu Chen. 2024. “Knowledge-Guided Heterogeneous Graph Convolutional Network for Aspect-Based Sentiment Analysis.” Electronics 13 (January). https://doi.org/10.3390/electronics13030517.
Song, Yang, Taesup Kim, Sebastian Nowozin, Stefano Ermon, and Nate Kushman. 2017. “PixelDefend: Leveraging Generative Models to Understand and Defend Against Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1710.10766.
Song, Yangqiu, Zhengdong Lu, Cane Wing-ki Leung, and Qiang Yang. 2013. “Collaborative Boosting for Activity Classification in Microblogs.” Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2487575.2487661.
Song, Yang, Ding Zhou, Jian Huang, Isaac Councill, Hongyuan Zha, and C. Giles. 2006. “Boosting the Feature Space: Text Classification for Unstructured Data on the Web.” Sixth International Conference on Data Mining (ICDM’06), December. https://doi.org/10.1109/icdm.2006.31.
Song, Yuhai, Zhong Cao, Kailun Wu, Ziang Yan, and Changshui Zhang. 2020. “Learning Fast Approximations of Sparse Nonlinear Regression.” arXiv. https://doi.org/10.48550/ARXIV.2010.13490.
Song, Zixing, Xiangli Yang, Zenglin Xu, and Irwin King. 2021. “Graph-Based Semi-Supervised Learning: A Comprehensive Review.” arXiv. https://doi.org/10.48550/ARXIV.2102.13303.
Sordoni, Alessandro, Philip Bachman, Adam Trischler, and Yoshua Bengio. 2016. “Iterative Alternating Neural Attention for Machine Reading.” arXiv. https://doi.org/10.48550/ARXIV.1606.02245.
Sordoni, Alessandro, Michel Galley, Michael Auli, Chris Brockett, Yangfeng Ji, Margaret Mitchell, Jian-Yun Nie, Jianfeng Gao, and Bill Dolan. 2015. “A Neural Network Approach to Context-Sensitive Generation of Conversational Responses.” arXiv. https://doi.org/10.48550/ARXIV.1506.06714.
Sordoni, Alessandro, Xingdi Yuan, Marc-Alexandre Côté, Matheus Pereira, Adam Trischler, Ziang Xiao, Arian Hosseini, Friederike Niedtner, and Nicolas Le Roux. 2023. “Joint Prompt Optimization of Stacked LLMs Using Variational Inference.” arXiv. https://doi.org/10.48550/ARXIV.2306.12509.
Sorokina, Daria, and Erick Cantu-Paz. 2016. “Amazon Search.” Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/2911451.2926725.
Sotoudeh, Matthew, and Sara S. Baghsorkhi. 2018. “DeepThin: A Self-Compressing Library for Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1802.06944.
South, L. F., C. J. Oates, A. Mira, and C. Drovandi. 2023. “Regularized Zero-Variance Control Variates.” Bayesian Analysis 18 (September). https://doi.org/10.1214/22-ba1328.
Sovdat, Blaz. 2014. “Updating Formulas and Algorithms for Computing Entropy and Gini Index from Time-Changing Data Streams.” arXiv. https://doi.org/10.48550/ARXIV.1403.6348.
Speer, Robert, Joshua Chin, Andrew Lin, Lance Nathan, and Sara Jewett. 2016. “Wordfreq: V1.5.1,” September. https://doi.org/10.5281/ZENODO.61937.
Speer, Robert, and Joanna Lowry-Duda. 2018. “Luminoso at SemEval-2018 Task 10: Distinguishing Attributes Using Text Corpora and Relational Knowledge.” Proceedings of The 12th International Workshop on Semantic Evaluation. https://doi.org/10.18653/v1/s18-1162.
Speiser, Jaime Lynn, Michael E. Miller, Janet Tooze, and Edward Ip. 2019. “A Comparison of Random Forest Variable Selection Methods for Classification Prediction Modeling.” Expert Systems with Applications 134 (November). https://doi.org/10.1016/j.eswa.2019.05.028.
Sperduti, A., and A. Starita. 1997. “Supervised Neural Networks for the Classification of Structures.” IEEE Transactions on Neural Networks 8 (May). https://doi.org/10.1109/72.572108.
Sponner, Max, Bernd Waschneck, and Akash Kumar. 2021. “Compiler Toolchains for Deep Learning Workloads on Embedded Platforms.” arXiv. https://doi.org/10.48550/ARXIV.2104.04576.
Springenberg, Jost Tobias, Alexey Dosovitskiy, Thomas Brox, and Martin Riedmiller. 2014. “Striving for Simplicity: The All Convolutional Net.” arXiv. https://doi.org/10.48550/ARXIV.1412.6806.
“Springer Handbook of Speech Processing.” 2008. Springer Handbooks. https://doi.org/10.1007/978-3-540-49127-9.
Srinivasan, Krishna, Karthik Raman, Jiecao Chen, Michael Bendersky, and Marc Najork. 2021. “WIT: Wikipedia-Based Image Text Dataset for Multimodal Multilingual Machine Learning.” Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3404835.3463257.
Srivastava, Rupesh Kumar, Klaus Greff, and Jürgen Schmidhuber. 2015. “Highway Networks.” arXiv. https://doi.org/10.48550/ARXIV.1505.00387.
Stachenfeld, Kimberly L., Matthew M. Botvinick, and Samuel J. Gershman. 2016. “The Hippocampus as a Predictive Map,” December. https://doi.org/10.1101/097170.
Staniak, Mateusz, and Przemysław Biecek. 2019. “Explanations of Model Predictions with Live and breakDown Packages.” The R Journal 10. https://doi.org/10.32614/rj-2018-072.
Stanley, Kenneth O., and Risto Miikkulainen. 2002. “Evolving Neural Networks Through Augmenting Topologies.” Evolutionary Computation 10 (June). https://doi.org/10.1162/106365602320169811.
Steinbrecher, Gregory R., Jonathan P. Olson, Dirk Englund, and Jacques Carolan. 2018. “Quantum Optical Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1808.10047.
Stevens, S. S. 1946. “On the Theory of Scales of Measurement.” Science 103 (June). https://doi.org/10.1126/science.103.2684.677.
Stich, Sebastian U., Jean-Baptiste Cordonnier, and Martin Jaggi. 2018. “Sparsified SGD with Memory.” arXiv. https://doi.org/10.48550/ARXIV.1809.07599.
Stoica, Ion, Dawn Song, Raluca Ada Popa, David Patterson, Michael W. Mahoney, Randy Katz, Anthony D. Joseph, et al. 2017. “A Berkeley View of Systems Challenges for AI.” arXiv. https://doi.org/10.48550/ARXIV.1712.05855.
Stokel-Walker, Chris. 2022. “AI Bot ChatGPT Writes Smart Essays — Should Professors Worry?” Nature, December. https://doi.org/10.1038/d41586-022-04397-7.
Strassen, V. 1964. “An Invariance Principle for the Law of the Iterated Logarithm.” Zeitschrift f�r Wahrscheinlichkeitstheorie Und Verwandte Gebiete 3. https://doi.org/10.1007/bf00534910.
Strobl, Carolin, Anne-Laure Boulesteix, Achim Zeileis, and Torsten Hothorn. 2006. “Bias in Random Forest Variable Importance Measures: Illustrations, Sources and a Solution.” Universitätsbibliothek Der Ludwig-Maximilians-Universität München. https://doi.org/10.5282/UBM/EPUB.1858.
Strobl, Eric V. 2023. “Root Causal Inference from Single Cell RNA Sequencing with the Negative Binomial.” arXiv. https://doi.org/10.48550/ARXIV.2307.05338.
Strohmann, Timo, Dominik Siemon, Bijan Khosrawi-Rad, and Susanne Robra-Bissantz. 2022. “Toward a Design Theory for Virtual Companionship.” Human–Computer Interaction 38 (July). https://doi.org/10.1080/07370024.2022.2084620.
Strub, Florian, Jeremie Mary, and Romaric Gaudel. 2016. “Hybrid Collaborative Filtering with Autoencoders.” arXiv. https://doi.org/10.48550/ARXIV.1603.00806.
Su, Jiahao, Jingling Li, Bobby Bhattacharjee, and Furong Huang. 2018. “Tensorial Neural Networks: Generalization of Neural Networks and Application to Model Compression.” arXiv. https://doi.org/10.48550/ARXIV.1805.10352.
Su, Jianlin, Yu Lu, Shengfeng Pan, Ahmed Murtadha, Bo Wen, and Yunfeng Liu. 2021. “RoFormer: Enhanced Transformer with Rotary Position Embedding.” arXiv. https://doi.org/10.48550/ARXIV.2104.09864.
Su, Jong-Chyi, Subhransu Maji, and Bharath Hariharan. 2019. “When Does Self-Supervision Improve Few-Shot Learning?” arXiv. https://doi.org/10.48550/ARXIV.1910.03560.
Su, Sen, Li Sun, Zhongbao Zhang, Gen Li, and Jielun Qu. 2018. “MASTER: Across Multiple Social Networks, Integrate Attribute and STructure Embedding for Reconciliation.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/537.
Su, Xiaoyuan, and Taghi M. Khoshgoftaar. 2009. “A Survey of Collaborative Filtering Techniques.” Advances in Artificial Intelligence 2009 (October). https://doi.org/10.1155/2009/421425.
Su, Ya, Youjian Zhao, Chenhao Niu, Rong Liu, Wei Sun, and Dan Pei. 2019. “Robust Anomaly Detection for Multivariate Time Series Through Stochastic Recurrent Neural Network.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330672.
Su, Yixuan, Tian Lan, Huayang Li, Jialu Xu, Yan Wang, and Deng Cai. 2023. “PandaGPT: One Model to Instruction-Follow Them All.” arXiv. https://doi.org/10.48550/ARXIV.2305.16355.
Su, Yu, Huan Sun, Brian Sadler, Mudhakar Srivatsa, Izzeddin Gur, Zenghui Yan, and Xifeng Yan. 2016. “On Generating Characteristic-Rich Question Sets for QA Evaluation.” Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/d16-1054.
Suarez, Joseph, Yilun Du, Phillip Isola, and Igor Mordatch. 2019. “Neural MMO: A Massively Multiagent Game Environment for Training and Evaluating Intelligent Agents.” arXiv. https://doi.org/10.48550/ARXIV.1903.00784.
Suau, Xavier, Luca Zappella, and Nicholas Apostoloff. 2018. “Filter Distillation for Network Compression.” arXiv. https://doi.org/10.48550/ARXIV.1807.10585.
Sugathadasa, Keet, Buddhi Ayesha, Nisansa de Silva, Amal Shehan Perera, Vindula Jayawardana, Dimuthu Lakmal, and Madhavi Perera. 2018. “Legal Document Retrieval Using Document Vector Embeddings and Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1805.10685.
Sukhbaatar, Sainbayar, Arthur Szlam, and Rob Fergus. 2016. “Learning Multiagent Communication with Backpropagation.” arXiv. https://doi.org/10.48550/ARXIV.1605.07736.
Sumner, Chris, Alison Byers, Rachel Boochever, and Gregory J. Park. 2012. “Predicting Dark Triad Personality Traits from Twitter Usage and a Linguistic Analysis of Tweets.” 2012 11th International Conference on Machine Learning and Applications, December. https://doi.org/10.1109/icmla.2012.218.
Sun, Baochen, Jiashi Feng, and Kate Saenko. 2015. “Return of Frustratingly Easy Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1511.05547.
Sun, Baochen, and Kate Saenko. 2016. “Deep CORAL: Correlation Alignment for Deep Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1607.01719.
Sun, Chong, Narasimhan Rampalli, Frank Yang, and AnHai Doan. 2014. “Chimera.” Proceedings of the VLDB Endowment 7 (August). https://doi.org/10.14778/2733004.2733024.
Sun, Gengxin, Sheng Bin, and Yixin Zhou. 2015. “Big Data Analytics of Multi-Relationship Online Social Network Based on Multi-Subnet Composited Complex Network.” International Journal of Database Theory and Application 8 (October). https://doi.org/10.14257/ijdta.2015.8.5.24.
Sun, Hao, Alihan Hüyük, and Mihaela van der Schaar. 2023. “Query-Dependent Prompt Evaluation and Optimization with Offline Inverse RL.” arXiv. https://doi.org/10.48550/ARXIV.2309.06553.
Sun, Hong, Xue Li, Yinchuan Xu, Youkow Homma, Qi Cao, Min Wu, Jian Jiao, and Denis Charles. 2023. “AutoHint: Automatic Prompt Optimization with Hint Generation.” arXiv. https://doi.org/10.48550/ARXIV.2307.07415.
Sun, Jiachen, Mark Ibrahim, Melissa Hall, Ivan Evtimov, Z. Morley Mao, Cristian Canton Ferrer, and Caner Hazirbas. 2023. “VPA: Fully Test-Time Visual Prompt Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.2309.15251.
Sun, Jingchen, Jiayu Qin, Zihao Lin, and Changyou Chen. 2023. “Prompt Tuning Based Adapter for Vision-Language Model Adaption.” arXiv. https://doi.org/10.48550/ARXIV.2303.15234.
Sun, Jingwei, Ang Li, Binghui Wang, Huanrui Yang, Hai Li, and Yiran Chen. 2020. “Provable Defense Against Privacy Leakage in Federated Learning from Representation Perspective.” arXiv. https://doi.org/10.48550/ARXIV.2012.06043.
Sun, Lichao, Yue Huang, Haoran Wang, Siyuan Wu, Qihui Zhang, Chujie Gao, Yixin Huang, et al. 2024. “TrustLLM: Trustworthiness in Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.05561.
Sun, Weiwei, Lingyong Yan, Xinyu Ma, Shuaiqiang Wang, Pengjie Ren, Zhumin Chen, Dawei Yin, and Zhaochun Ren. 2023. “Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agents.” arXiv. https://doi.org/10.48550/ARXIV.2304.09542.
Sun, Xiangguo, Jiawen Zhang, Xixi Wu, Hong Cheng, Yun Xiong, and Jia Li. 2023. “Graph Prompt Learning: A Comprehensive Survey and Beyond.” arXiv. https://doi.org/10.48550/ARXIV.2311.16534.
Sun, Xu, Xuancheng Ren, Shuming Ma, Bingzhen Wei, Wei Li, Jingjing Xu, Houfeng Wang, and Yi Zhang. 2020. “Training Simplification and Model Simplification for Deep Learning : A Minimal Effort Back Propagation Method.” IEEE Transactions on Knowledge and Data Engineering 32 (February). https://doi.org/10.1109/tkde.2018.2883613.
Sun, Yanmin, Mohamed Kamel, and Yang Wang. 2006. “Boosting for Learning Multiple Classes with Imbalanced Class Distribution.” Sixth International Conference on Data Mining (ICDM’06), December. https://doi.org/10.1109/icdm.2006.29.
Sun, Yi, Abel Valente, Sijia Liu, and Dakuo Wang. 2021. “Preserve, Promote, or Attack? GNN Explanation via Topology Perturbation.” arXiv. https://doi.org/10.48550/ARXIV.2103.13944.
Sun, Yueming, and Yi Zhang. 2018. “Conversational Recommender System.” arXiv. https://doi.org/10.48550/ARXIV.1806.03277.
Sun, Yutao, Li Dong, Shaohan Huang, Shuming Ma, Yuqing Xia, Jilong Xue, Jianyong Wang, and Furu Wei. 2023. “Retentive Network: A Successor to Transformer for Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2307.08621.
Sun, Zehua, Qiuhong Ke, Hossein Rahmani, Mohammed Bennamoun, Gang Wang, and Jun Liu. 2022. “Human Action Recognition from Various Data Modalities: A Review.” IEEE Transactions on Pattern Analysis and Machine Intelligence. https://doi.org/10.1109/tpami.2022.3183112.
Sun, Zequn, Wei Hu, and Chengkai Li. 2017. “Cross-Lingual Entity Alignment via Joint Attribute-Preserving Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1708.05045.
Sun, Zhiqing, Zhi-Hong Deng, Jian-Yun Nie, and Jian Tang. 2019. “RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space.” arXiv. https://doi.org/10.48550/ARXIV.1902.10197.
Sundararajan, Mukund, Ankur Taly, and Qiqi Yan. 2017. “Axiomatic Attribution for Deep Networks.” arXiv. https://doi.org/10.48550/ARXIV.1703.01365.
Suo, Xuchen. 2024. “Signed-Prompt: A New Approach to Prevent Prompt Injection Attacks Against LLM-Integrated Applications.” arXiv. https://doi.org/10.48550/ARXIV.2401.07612.
Supic, Lazar, Rawan Naous, Ranko Sredojevic, Aleksandra Faust, and Vladimir Stojanovic. 2018. “MPDCompress - Matrix Permutation Decomposition Algorithm for Deep Neural Network Compression.” arXiv. https://doi.org/10.48550/ARXIV.1805.12085.
Supriya, S., S. Siuly, and Y. Zhang. 2016. “Automatic Epilepsy Detection from EEG Introducing a New Edge Weight Method in the Complex Network.” Electronics Letters 52 (August). https://doi.org/10.1049/el.2016.1992.
Surís, Dídac, Sachit Menon, and Carl Vondrick. 2023. “ViperGPT: Visual Inference via Python Execution for Reasoning.” arXiv. https://doi.org/10.48550/ARXIV.2303.08128.
Susnjak, Teo. 2022. “ChatGPT: The End of Online Exam Integrity?” arXiv. https://doi.org/10.48550/ARXIV.2212.09292.
———. 2023. “Applying BERT and ChatGPT for Sentiment Analysis of Lyme Disease in Scientific Literature.” arXiv. https://doi.org/10.48550/ARXIV.2302.06474.
Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. “Sequence to Sequence Learning with Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1409.3215.
Sutter, David. 2018. “Approximate Quantum Markov Chains.” arXiv. https://doi.org/10.48550/ARXIV.1802.05477.
Sverrisson, Freyr, Jean Feydy, Bruno E. Correia, and Michael M. Bronstein. 2020. “Fast End-to-End Learning on Protein Surfaces,” December. https://doi.org/10.1101/2020.12.28.424589.
Syafrudin, Muhammad, Ganjar Alfian, Norma Fitriyani, and Jongtae Rhee. 2018. “Performance Analysis of IoT-Based Sensor, Big Data Processing, and Machine Learning Model for Real-Time Monitoring System in Automotive Manufacturing.” Sensors 18 (September). https://doi.org/10.3390/s18092946.
Szegedy, Christian, Sergey Ioffe, Vincent Vanhoucke, and Alex Alemi. 2016. “Inception-V4, Inception-ResNet and the Impact of Residual Connections on Learning.” arXiv. https://doi.org/10.48550/ARXIV.1602.07261.
Szegedy, Christian, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. 2014. “Going Deeper with Convolutions.” arXiv. https://doi.org/10.48550/ARXIV.1409.4842.
Szegedy, Christian, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. 2016. “Rethinking the Inception Architecture for Computer Vision.” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr.2016.308.
Szegedy, Christian, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2013. “Intriguing Properties of Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1312.6199.
Taha, Ahmed, Abhinav Shrivastava, and Larry Davis. 2021. “Knowledge Evolution in Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2103.05152.
Taherian, Hassan. 2016. “End-to-End Attention-Based Distant Speech Recognition with Highway LSTM.” arXiv. https://doi.org/10.48550/ARXIV.1610.05361.
Tai, Cheng, Tong Xiao, Yi Zhang, Xiaogang Wang, and Weinan E. 2015. “Convolutional Neural Networks with Low-Rank Regularization.” arXiv. https://doi.org/10.48550/ARXIV.1511.06067.
Tai, Kai Sheng, Richard Socher, and Christopher D. Manning. 2015. “Improved Semantic Representations from Tree-Structured Long Short-Term Memory Networks.” arXiv. https://doi.org/10.48550/ARXIV.1503.00075.
Taieb, Souhaib Ben, and Bonsoo Koo. 2019. “Regularized Regression for Hierarchical Forecasting Without Unbiasedness Conditions.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330976.
Taigman, Yaniv, Adam Polyak, and Lior Wolf. 2016. “Unsupervised Cross-Domain Image Generation.” arXiv. https://doi.org/10.48550/ARXIV.1611.02200.
Takanobu, Ryuichi, Tao Zhuang, Minlie Huang, Jun Feng, Haihong Tang, and Bo Zheng. 2019. “Aggregating e-Commerce Search Results from Heterogeneous Sources via Hierarchical Reinforcement Learning.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313455.
Tam, Nguyen Thanh, Matthias Weidlich, Bolong Zheng, Hongzhi Yin, Nguyen Quoc Viet Hung, and Bela Stantic. 2019. “From Anomaly Detection to Rumour Detection Using Data Streams of Social Platforms.” Proceedings of the VLDB Endowment 12 (May). https://doi.org/10.14778/3329772.3329778.
Tampuu, Ardi, Tambet Matiisen, Dorian Kodelja, Ilya Kuzovkin, Kristjan Korjus, Juhan Aru, Jaan Aru, and Raul Vicente. 2015. “Multiagent Cooperation and Competition with Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1511.08779.
Tan, Chenhao, Lillian Lee, and Bo Pang. 2014. “The Effect of Wording on Message Propagation: Topic- and Author-Controlled Natural Experiments on Twitter.” arXiv. https://doi.org/10.48550/ARXIV.1405.1438.
Tan, Chenhao, Lillian Lee, Jie Tang, Long Jiang, Ming Zhou, and Ping Li. 2011. “User-Level Sentiment Analysis Incorporating Social Networks.” arXiv. https://doi.org/10.48550/ARXIV.1109.6018.
Tan, Chuanqi, Fuchun Sun, Tao Kong, Wenchang Zhang, Chao Yang, and Chunfang Liu. 2018. “A Survey on Deep Transfer Learning.” arXiv. https://doi.org/10.48550/ARXIV.1808.01974.
Tan, Hao, and Mohit Bansal. 2019. “LXMERT: Learning Cross-Modality Encoder Representations from Transformers.” arXiv. https://doi.org/10.48550/ARXIV.1908.07490.
Tan, Mingkui, Ivor W. Tsang, and Li Wang. 2013. “Minimax Sparse Logistic Regression for Very High-Dimensional Feature Selection.” IEEE Transactions on Neural Networks and Learning Systems 24 (October). https://doi.org/10.1109/tnnls.2013.2263427.
Tan, Sarah, Giles Hooker, Paul Koch, Albert Gordo, and Rich Caruana. 2023. “Considerations When Learning Additive Explanations for Black-Box Models.” Machine Learning 112 (June). https://doi.org/10.1007/s10994-023-06335-8.
Tanaka, Paulo, Sameet Sapra, and Nikolay Laptev. 2020. “Scalable Data Classification for Security and Privacy.” arXiv. https://doi.org/10.48550/ARXIV.2006.14109.
Tang, Haotian, Zhijian Liu, Shengyu Zhao, Yujun Lin, Ji Lin, Hanrui Wang, and Song Han. 2020. “Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution.” arXiv. https://doi.org/10.48550/ARXIV.2007.16100.
Tang, Jialin, Slim Soua, Cristinel Mares, and Tat-Hean Gan. 2017. “A Pattern Recognition Approach to Acoustic Emission Data Originating from Fatigue of Wind Turbine Blades.” Sensors 17 (November). https://doi.org/10.3390/s17112507.
Tang, Jiaxi, and Ke Wang. 2018. “Ranking Distillation.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220021.
Tang, Ruixiang, Xiaotian Han, Xiaoqian Jiang, and Xia Hu. 2023. “Does Synthetic Data Generation of LLMs Help Clinical Text Mining?” arXiv. https://doi.org/10.48550/ARXIV.2303.04360.
Tang, Xianfeng, Yandong Li, Yiwei Sun, Huaxiu Yao, Prasenjit Mitra, and Suhang Wang. 2020. “Transferring Robustness for Graph Neural Network Against Poisoning Attacks.” Proceedings of the 13th International Conference on Web Search and Data Mining, January. https://doi.org/10.1145/3336191.3371851.
Tang, Xianfeng, Huaxiu Yao, Yiwei Sun, Yiqi Wang, Jiliang Tang, Charu Aggarwal, Prasenjit Mitra, and Suhang Wang. 2020. “Investigating and Mitigating Degree-Related Biases in Graph Convoltuional Networks.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3411872.
Tanwisuth, Korawat, Shujian Zhang, Huangjie Zheng, Pengcheng He, and Mingyuan Zhou. 2023. “POUF: Prompt-Oriented Unsupervised Fine-Tuning for Large Pre-Trained Models.” arXiv. https://doi.org/10.48550/ARXIV.2305.00350.
Tao, Renshuai, Hainan Li, Tianbo Wang, Yanlu Wei, Yifu Ding, Bowei Jin, Hongping Zhi, Xianglong Liu, and Aishan Liu. 2022. “Exploring Endogenous Shift for Cross-Domain Detection: A Large-Scale Benchmark and Perturbation Suppression Network.” 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52688.2022.02051.
Tarr, Alexander, June Hwang, and Kosuke Imai. 2022. “Automated Coding of Political Campaign Advertisement Videos: An Empirical Validation Study.” Political Analysis 31 (November). https://doi.org/10.1017/pan.2022.26.
Tassa, Yuval, Nicolas Mansard, and Emo Todorov. 2014. “Control-Limited Differential Dynamic Programming.” 2014 IEEE International Conference on Robotics and Automation (ICRA), May. https://doi.org/10.1109/icra.2014.6907001.
Taylor, Sean J, and Benjamin Letham. 2017. “Forecasting at Scale,” September. https://doi.org/10.7287/peerj.preprints.3190v2.
Teebagy, Sean, Lauren Colwell, Emma Wood, Antonio Yaghy, and Misha Faustina. 2023. “Improved Performance of ChatGPT-4 on the OKAP Exam: A Comparative Study with ChatGPT-3.5,” April. https://doi.org/10.1101/2023.04.03.23287957.
Teh, Yee Whye, and Michael I. Jordan. 2010. “Hierarchical Bayesian Nonparametric Models with Applications.” Bayesian Nonparametrics, April. https://doi.org/10.1017/cbo9780511802478.006.
Teney, Damien, Lingqiao Liu, and Anton van den Hengel. 2016. “Graph-Structured Representations for Visual Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1609.05600.
Tenney, Ian, Patrick Xia, Berlin Chen, Alex Wang, Adam Poliak, R Thomas McCoy, Najoung Kim, et al. 2019. “What Do You Learn from Context? Probing for Sentence Structure in Contextualized Word Representations.” arXiv. https://doi.org/10.48550/ARXIV.1905.06316.
Teubner, Timm, Christoph M. Flath, Christof Weinhardt, Wil van der Aalst, and Oliver Hinz. 2023. “Welcome to the Era of ChatGPT Et Al.” Business &Amp; Information Systems Engineering 65 (March). https://doi.org/10.1007/s12599-023-00795-x.
Theodosiou, Filotas, and Nikolaos Kourentzes. 2021. “Forecasting with Deep Temporal Hierarchies.” SSRN Electronic Journal. https://doi.org/10.2139/ssrn.3918315.
Thoma, Martin. 2017. “Analysis and Optimization of Convolutional Neural Network Architectures.” arXiv. https://doi.org/10.48550/ARXIV.1707.09725.
Thorne, James, Andreas Vlachos, Christos Christodoulopoulos, and Arpit Mittal. 2018. “FEVER: A Large-Scale Dataset for Fact Extraction and VERification.” arXiv. https://doi.org/10.48550/ARXIV.1803.05355.
Thorne, James, Majid Yazdani, Marzieh Saeidi, Fabrizio Silvestri, Sebastian Riedel, and Alon Halevy. 2020. “Neural Databases.” arXiv. https://doi.org/10.48550/ARXIV.2010.06973.
Thys, Simen, Wiebe Van Ranst, and Toon Goedemé. 2019. “Fooling Automated Surveillance Cameras: Adversarial Patches to Attack Person Detection.” arXiv. https://doi.org/10.48550/ARXIV.1904.08653.
Tian, Haoye, Weiqi Lu, Tsz On Li, Xunzhu Tang, Shing-Chi Cheung, Jacques Klein, and Tegawendé F. Bissyandé. 2023. “Is ChatGPT the Ultimate Programming Assistant – How Far Is It?” arXiv. https://doi.org/10.48550/ARXIV.2304.11938.
Tian, Junjiao, Xiaoliang Dai, Chih-Yao Ma, Zecheng He, Yen-Cheng Liu, and Zsolt Kira. 2023. “Trainable Projected Gradient Method for Robust Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2303.10720.
Tian, Qing, Tal Arbel, and James J. Clark. 2021. “Task Dependent Deep LDA Pruning of Neural Networks.” Computer Vision and Image Understanding 203 (February). https://doi.org/10.1016/j.cviu.2020.103154.
Tian, Shubo, Qiao Jin, Lana Yeganova, Po-Ting Lai, Qingqing Zhu, Xiuying Chen, Yifan Yang, et al. 2023. “Opportunities and Challenges for ChatGPT and Large Language Models in Biomedicine and Health.” Briefings in Bioinformatics 25 (November). https://doi.org/10.1093/bib/bbad493.
Tian, Xinyu, Shu Zou, Zhaoyuan Yang, and Jing Zhang. 2023. “ArGue: Attribute-Guided Prompt Tuning for Vision-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2311.16494.
Tian, Yue, Guanjun Liu, Jiacun Wang, and Mengchu Zhou. 2023. “Transaction Fraud Detection via an Adaptive Graph Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.2307.05633.
Tian, Zhengkun, Jiangyan Yi, Jianhua Tao, Ye Bai, and Zhengqi Wen. 2019. “Self-Attention Transducers for End-to-End Speech Recognition.” Interspeech 2019, September. https://doi.org/10.21437/interspeech.2019-2203.
Tichavsky, Petr, Anh Huy Phan, and Andrzej Cichocki. 2014. “Non-Orthogonal Tensor Diagonalization.” arXiv. https://doi.org/10.48550/ARXIV.1402.1673.
Tifrea, Alexandru, Gary Bécigneul, and Octavian-Eugen Ganea. 2018. “Poincaré GloVe: Hyperbolic Word Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1810.06546.
Tillo, Desiree, and Timothy R Hughes. 2009. “G+c Content Dominates Intrinsic Nucleosome Occupancy.” BMC Bioinformatics 10 (December). https://doi.org/10.1186/1471-2105-10-442.
Tindel, S., C. A. Tudor, and F. Viens. 2003. “Stochastic Evolution Equations with Fractional Brownian Motion.” Probability Theory and Related Fields 127 (July). https://doi.org/10.1007/s00440-003-0282-2.
Ting, Kai Ming, Bi-Cun Xu, Takashi Washio, and Zhi-Hua Zhou. 2020. “Isolation Distributional Kernel: A New Tool for Point &Amp; Group Anomaly Detection.” arXiv. https://doi.org/10.48550/ARXIV.2009.12196.
Toderici, George, Damien Vincent, Nick Johnston, Sung Jin Hwang, David Minnen, Joel Shor, and Michele Covell. 2016. “Full Resolution Image Compression with Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1608.05148.
Togninalli, Matteo, Elisabetta Ghisu, Felipe Llinares-López, Bastian Rieck, and Karsten Borgwardt. 2019. “Wasserstein Weisfeiler-Lehman Graph Kernels.” arXiv. https://doi.org/10.48550/ARXIV.1906.01277.
Tokdar, Surya T., and Joseph B. Kadane. 2012. “Simultaneous Linear Quantile Regression: A Semiparametric Bayesian Approach.” Bayesian Analysis 7 (March). https://doi.org/10.1214/12-ba702.
Tomasi, Federico, Rishabh Mehrotra, Aasish Pappu, Judith Bütepage, Brian Brost, Hugo Galvão, and Mounia Lalmas. 2020. “Query Understanding for Surfacing Under-Served Music Content.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412741.
Tong, Mukun, Charles Dawson, and Chuchu Fan. 2022. “Enforcing Safety for Vision-Based Controllers via Control Barrier Functions and Neural Radiance Fields.” arXiv. https://doi.org/10.48550/ARXIV.2209.12266.
Tong, Zhiqiang, and Gouhei Tanaka. 2015. “A Pruning Method Based on Weight Variation Information for Feedforward Neural Networks.” IFAC-PapersOnLine 48. https://doi.org/10.1016/j.ifacol.2015.11.040.
Törnberg, Petter. 2023. “ChatGPT-4 Outperforms Experts and Crowd Workers in Annotating Political Twitter Messages with Zero-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.2304.06588.
Toutanova, Kristina, Danqi Chen, Patrick Pantel, Hoifung Poon, Pallavi Choudhury, and Michael Gamon. 2015. “Representing Text for Joint Embedding of Text and Knowledge Bases.” Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. https://doi.org/10.18653/v1/d15-1174.
Touvron, Hugo, Matthieu Cord, Alaaeldin El-Nouby, Jakob Verbeek, and Hervé Jégou. 2022. “Three Things Everyone Should Know about Vision Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2203.09795.
Touvron, Hugo, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, et al. 2023. “Llama 2: Open Foundation and Fine-Tuned Chat Models.” arXiv. https://doi.org/10.48550/ARXIV.2307.09288.
Towell, Geoffrey G., and Jude W. Shavlik. 1993. “Extracting Refined Rules from Knowledge-Based Neural Networks.” Machine Learning 13 (October). https://doi.org/10.1007/bf00993103.
Tramèr, Florian, Alexey Kurakin, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick McDaniel. 2017. “Ensemble Adversarial Training: Attacks and Defenses.” arXiv. https://doi.org/10.48550/ARXIV.1705.07204.
Tramèr, Florian, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick McDaniel. 2017. “The Space of Transferable Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.1704.03453.
Tramèr, Florian, Fan Zhang, Ari Juels, Michael K. Reiter, and Thomas Ristenpart. 2016. “Stealing Machine Learning Models via Prediction APIs.” arXiv. https://doi.org/10.48550/ARXIV.1609.02943.
Tran, Nhan, Ben Kreis, Javier Duarte, Duc Hoang, Adrian Alan Pol, Dejan Golubovic, Paolo Cretaro, and Petr Zejdl. 2021. “Fastmachinelearning/Hls4ml: Bartsia,” March. https://doi.org/10.5281/ZENODO.4585796.
Tresp, Volker. 2000. “A Bayesian Committee Machine.” Neural Computation 12 (November). https://doi.org/10.1162/089976600300014908.
Triantafillou, Eleni, Tyler Zhu, Vincent Dumoulin, Pascal Lamblin, Utku Evci, Kelvin Xu, Ross Goroshin, et al. 2019. “Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples.” arXiv. https://doi.org/10.48550/ARXIV.1903.03096.
Triantafyllopoulos, Andreas, Anastasia Semertzidou, Meishu Song, Florian B. Pokorny, and Björn W. Schuller. 2022. “COVYT: Introducing the Coronavirus YouTube and TikTok Speech Dataset Featuring the Same Speakers with and Without Infection.” arXiv. https://doi.org/10.48550/ARXIV.2206.11045.
Trischler, Adam, Tong Wang, Xingdi Yuan, Justin Harris, Alessandro Sordoni, Philip Bachman, and Kaheer Suleman. 2016. “NewsQA: A Machine Comprehension Dataset.” arXiv. https://doi.org/10.48550/ARXIV.1611.09830.
Trivedi, Rakshit, Hanjun Dai, Yichen Wang, and Le Song. 2017. “Know-Evolve: Deep Temporal Reasoning for Dynamic Knowledge Graphs.” arXiv. https://doi.org/10.48550/ARXIV.1705.05742.
Trosten, Daniel J., Sigurd Løkse, Robert Jenssen, and Michael C. Kampffmeyer. 2023. “On the Effects of Self-Supervision and Contrastive Alignment in Deep Multi-View Clustering.” arXiv. https://doi.org/10.48550/ARXIV.2303.09877.
Trubowitz, Peter, and Kohei Watanabe. 2021. “The Geopolitical Threat Index: A Text-Based Computational Approach to Identifying Foreign Threats.” International Studies Quarterly 65 (May). https://doi.org/10.1093/isq/sqab029.
Tsai, Henry, Jason Riesa, Melvin Johnson, Naveen Arivazhagan, Xin Li, and Amelia Archer. 2019. “Small and Practical BERT Models for Sequence Labeling.” arXiv. https://doi.org/10.48550/ARXIV.1909.00100.
Tu, Cunchao, Xiangkai Zeng, Hao Wang, Zhengyan Zhang, Zhiyuan Liu, Maosong Sun, Bo Zhang, and Leyu Lin. 2016. “A Unified Framework for Community Detection and Network Representation Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.06645.
Tu, Zhaopeng, Zhengdong Lu, Yang Liu, Xiaohua Liu, and Hang Li. 2016. “Modeling Coverage for Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1601.04811.
Tuli, Shikhar, Bhishma Dedhia, Shreshth Tuli, and Niraj K. Jha. 2022. “FlexiBERT: Are Current Transformer Architectures Too Homogeneous and Rigid?” arXiv. https://doi.org/10.48550/ARXIV.2205.11656.
Tung, Frederick, Srikanth Muralidharan, and Greg Mori. 2017. “Fine-Pruning: Joint Fine-Tuning and Compression of a Convolutional Network with Bayesian Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1707.09102.
Turner, Richard Eric, and Maneesh Sahani. 2011. “Two Problems with Variational Expectation Maximisation for Time Series Models.” Bayesian Time Series Models, August. https://doi.org/10.1017/cbo9780511984679.006.
Turrisi, Rosanna, Rémi Flamary, Alain Rakotomamonjy, and Massimiliano Pontil. 2020. “Multi-Source Domain Adaptation via Weighted Joint Distributions Optimal Transport.” arXiv. https://doi.org/10.48550/ARXIV.2006.12938.
Tuson, Matthew, Matthew Yap, Mei Ruu Kok, Bryan Boruff, Kevin Murray, Alistair Vickery, Berwin A. Turlach, and David Whyatt. 2020. “Overcoming Inefficiencies Arising Due to the Impact of the Modifiable Areal Unit Problem on Single-Aggregation Disease Maps.” International Journal of Health Geographics 19 (October). https://doi.org/10.1186/s12942-020-00236-y.
Tyree, Stephen, Kilian Q. Weinberger, Kunal Agrawal, and Jennifer Paykin. 2011. “Parallel Boosted Regression Trees for Web Search Ranking.” Proceedings of the 20th International Conference on World Wide Web, March. https://doi.org/10.1145/1963405.1963461.
Tzagkarakis, Christos, Pavlos Charalampidis, Stylianos Roubakis, Alexandros Fragkiadakis, and Sotiris Ioannidis. 2022. “Evaluating Short-Term Forecasting of Multiple Time Series in IoT Environments.” arXiv. https://doi.org/10.48550/ARXIV.2206.07784.
Tzeng, Eric, Judy Hoffman, Ning Zhang, Kate Saenko, and Trevor Darrell. 2014. “Deep Domain Confusion: Maximizing for Domain Invariance.” arXiv. https://doi.org/10.48550/ARXIV.1412.3474.
Uc-Cetina, Víctor, Nicolás Navarro-Guerrero, Anabel Martin-Gonzalez, Cornelius Weber, and Stefan Wermter. 2022. “Survey on Reinforcement Learning for Language Processing.” Artificial Intelligence Review 56 (June). https://doi.org/10.1007/s10462-022-10205-5.
Ulyanov, Dmitry, Andrea Vedaldi, and Victor Lempitsky. 2016. “Instance Normalization: The Missing Ingredient for Fast Stylization.” arXiv. https://doi.org/10.48550/ARXIV.1607.08022.
Unterkalmsteiner, Michael, Pekka Abrahamsson, Wang XiaoFeng, Anh Nguyen-Duc, Syed Shah, Sohaib Shahid Bajwa, Guido H. Baltes, et al. 2016. “Software Startups - a Research Agenda.” E-Informatica Vol. X. https://doi.org/10.5277/E-INF160105.
Urban, Gregor, Krzysztof J. Geras, Samira Ebrahimi Kahou, Ozlem Aslan, Shengjie Wang, Rich Caruana, Abdelrahman Mohamed, Matthai Philipose, and Matt Richardson. 2016. “Do Deep Convolutional Nets Really Need to Be Deep and Convolutional?” arXiv. https://doi.org/10.48550/ARXIV.1603.05691.
“User Modeling, Adaption and Personalization.” 2011. Lecture Notes in Computer Science. https://doi.org/10.1007/978-3-642-22362-4.
Uwents, Werner, Gabriele Monfardini, Hendrik Blockeel, Marco Gori, and Franco Scarselli. 2010. “Neural Networks for Relational Learning: An Experimental Comparison.” Machine Learning 82 (July). https://doi.org/10.1007/s10994-010-5196-5.
“UZH in BioNLP 2013.” 2013. Association for Computational Linguistics. https://doi.org/10.5167/UZH-91884.
Vahab, Mohammad, Ehsan Haghighat, Maryam Khaleghi, and Nasser Khalili. 2021. “A Physics Informed Neural Network Approach to Solution and Identification of Biharmonic Equations of Elasticity.” arXiv. https://doi.org/10.48550/ARXIV.2108.07243.
Vaidya, Aatman, Arnav Arora, Aditya Joshi, and Tarunima Prabhakar. 2024. “Overview of the 2023 ICON Shared Task on Gendered Abuse Detection in Indic Languages.” arXiv. https://doi.org/10.48550/ARXIV.2401.03677.
Valdes, Gilmer, José Marcio Luna, Efstathios D. Gennatas, Lyle H. Ungar, Eric Eaton, Eric S. Diffenderfer, Shane T. Jensen, Charles B. Simone, Jerome H. Friedman, and Timothy D. Solberg. 2020. “Reply to Nock and Nielsen: On the Work of Nock and Nielsen and Its Relationship to the Additive Tree.” Proceedings of the National Academy of Sciences 117 (April). https://doi.org/10.1073/pnas.2002399117.
Valmeekam, Karthik, Matthew Marquez, Alberto Olmo, Sarath Sreedharan, and Subbarao Kambhampati. 2022. “PlanBench: An Extensible Benchmark for Evaluating Large Language Models on Planning and Reasoning about Change.” arXiv. https://doi.org/10.48550/ARXIV.2206.10498.
Valmeekam, Karthik, Sarath Sreedharan, Matthew Marquez, Alberto Olmo, and Subbarao Kambhampati. 2023. “On the Planning Abilities of Large Language Models (a Critical Investigation with a Proposed Benchmark).” arXiv. https://doi.org/10.48550/ARXIV.2302.06706.
Vapnik, V. N., and A. Ya. Chervonenkis. 1971. “On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities.” Theory of Probability &Amp; Its Applications 16 (January). https://doi.org/10.1137/1116025.
“Variable Length Markov Chains: Methodology, Computing and Software.” 2002. ETH Zurich. https://doi.org/10.3929/ETHZ-A-004310029.
Varley, Jacob, Chad DeChant, Adam Richardson, Joaquín Ruales, and Peter Allen. 2016. “Shape Completion Enabled Robotic Grasping.” arXiv. https://doi.org/10.48550/ARXIV.1609.08546.
Varma, Paroma, Frederic Sala, Ann He, Alexander Ratner, and Christopher Ré. 2019. “Learning Dependency Structures for Weak Supervision Models.” arXiv. https://doi.org/10.48550/ARXIV.1903.05844.
Vasek, Marie. 2014. “There’s No Free Lunch, Even Using Bitcoin: Tracking the Popularity and Profits of Virtual Currency Scams.” :unav. https://doi.org/10.7910/DVN/28561.
Vashishth, Shikhar, Prince Jain, and Partha Talukdar. 2018. “CESI.” Proceedings of the 2018 World Wide Web Conference on World Wide Web - WWW ’18. https://doi.org/10.1145/3178876.3186030.
Vashishth, Shikhar, Soumya Sanyal, Vikram Nitin, Nilesh Agrawal, and Partha Talukdar. 2019. “InteractE: Improving Convolution-Based Knowledge Graph Embeddings by Increasing Feature Interactions.” arXiv. https://doi.org/10.48550/ARXIV.1911.00219.
Vasile, Flavian, Damien Lefortier, and Olivier Chapelle. 2016. “Cost-Sensitive Learning for Utility Optimization in Online Advertising Auctions.” arXiv. https://doi.org/10.48550/ARXIV.1603.03713.
Vasiloudis, Theodore, Foteini Beligianni, and Gianmarco De Francisci Morales. 2017. “BoostVHT.” Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3132847.3132974.
Vasquez, Sean, and Mike Lewis. 2019. “MelNet: A Generative Model for Audio in the Frequency Domain.” arXiv. https://doi.org/10.48550/ARXIV.1906.01083.
Vasudevan, Sriram, and Krishnaram Kenthapadi. 2020. “LiFT.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412705.
Vedaldi, Andrea, and Karel Lenc. 2014. “MatConvNet - Convolutional Neural Networks for MATLAB.” arXiv. https://doi.org/10.48550/ARXIV.1412.4564.
Vedantam, Ramakrishna, C. Lawrence Zitnick, and Devi Parikh. 2014. “CIDEr: Consensus-Based Image Description Evaluation.” arXiv. https://doi.org/10.48550/ARXIV.1411.5726.
Vehtari, Aki, Andrew Gelman, Daniel Simpson, Bob Carpenter, and Paul-Christian Bürkner. 2021. “Rank-Normalization, Folding, and Localization: An Improved Rˆ for Assessing Convergence of MCMC (with Discussion).” Bayesian Analysis 16 (June). https://doi.org/10.1214/20-ba1221.
Vehtari, Aki, Daniel Simpson, Andrew Gelman, Yuling Yao, and Jonah Gabry. 2015. “Pareto Smoothed Importance Sampling.” arXiv. https://doi.org/10.48550/ARXIV.1507.02646.
Veličković, Petar, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. 2017. “Graph Attention Networks,” October. http://arxiv.org/abs/1710.10903v3.
Venieris, Stylianos I., and Christos-Savvas Bouganis. 2017. “fpgaConvNet: A Toolflow for Mapping Diverse Convolutional Neural Networks on Embedded FPGAs.” arXiv. https://doi.org/10.48550/ARXIV.1711.08740.
Venkataramanaiah, Shreyas Kolala, Yufei Ma, Shihui Yin, Eriko Nurvithadhi, Aravind Dasu, Yu Cao, and Jae-sun Seo. 2019. “Automatic Compiler Based FPGA Accelerator for CNN Training.” arXiv. https://doi.org/10.48550/ARXIV.1908.06724.
Ventura, Dan, and Tony Martinez. 1998. “Quantum Associative Memory.” arXiv. https://doi.org/10.48550/ARXIV.QUANT-PH/9807053.
Ventura, Sebastián, Cristóbal Romero, Amelia Zafra, José A. Delgado, and César Hervás. 2007. “JCLEC: A Java Framework for Evolutionary Computation.” Soft Computing 12 (April). https://doi.org/10.1007/s00500-007-0172-0.
Verga, Pat, Haitian Sun, Livio Baldini Soares, and William W. Cohen. 2020. “Facts as Experts: Adaptable and Interpretable Neural Memory over Symbolic Knowledge.” arXiv. https://doi.org/10.48550/ARXIV.2007.00849.
Veronis, Jean, and Nancy M. Ide. 1990. “Word Sense Disambiguation with Very Large Neural Networks Extracted from Machine Readable Dictionaries.” Proceedings of the 13th Conference on Computational Linguistics -. https://doi.org/10.3115/997939.998006.
Vig, Jesse. 2019. “A Multiscale Visualization of Attention in the Transformer Model.” arXiv. https://doi.org/10.48550/ARXIV.1906.05714.
Vilnis, Luke, and Andrew McCallum. 2014. “Word Representations via Gaussian Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1412.6623.
Vinyals, Oriol, Charles Blundell, Timothy Lillicrap, Koray Kavukcuoglu, and Daan Wierstra. 2016. “Matching Networks for One Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.04080.
Vinyals, Oriol, Lukasz Kaiser, Terry Koo, Slav Petrov, Ilya Sutskever, and Geoffrey Hinton. 2014. “Grammar as a Foreign Language.” arXiv. https://doi.org/10.48550/ARXIV.1412.7449.
Vinyals, Oriol, and Quoc Le. 2015. “A Neural Conversational Model.” arXiv. https://doi.org/10.48550/ARXIV.1506.05869.
Vinyals, Oriol, Alexander Toshev, Samy Bengio, and Dumitru Erhan. 2014. “Show and Tell: A Neural Image Caption Generator.” arXiv. https://doi.org/10.48550/ARXIV.1411.4555.
Virani, Alim, Jay Baxter, Dan Shiebler, Philip Gautier, Shivam Verma, Yan Xia, Apoorv Sharma, Sumit Binnani, Linlin Chen, and Chenguang Yu. 2021. “Lessons Learned Addressing Dataset Bias in Model-Based Candidate Generation at Twitter.” arXiv. https://doi.org/10.48550/ARXIV.2105.09293.
Vishwanathan, S. V. N., Nicol N. Schraudolph, Mark W. Schmidt, and Kevin P. Murphy. 2006. “Accelerated Training of Conditional Random Fields with Stochastic Gradient Methods.” Proceedings of the 23rd International Conference on Machine Learning - ICML ’06. https://doi.org/10.1145/1143844.1143966.
Vogel, David S., Ognian Asparouhov, and Tobias Scheffer. 2007. “Scalable Look-Ahead Linear Regression Trees.” Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/1281192.1281273.
Voita, Elena, David Talbot, Fedor Moiseev, Rico Sennrich, and Ivan Titov. 2019. “Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned.” Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/p19-1580.
Vollmer, Sebastian, Bilal A Mateen, Gergo Bohner, Franz J Király, Rayid Ghani, Pall Jonsson, Sarah Cumbers, et al. 2020. “Machine Learning and Artificial Intelligence Research for Patient Benefit: 20 Critical Questions on Transparency, Replicability, Ethics, and Effectiveness.” BMJ, March. https://doi.org/10.1136/bmj.l6927.
Vovk, Vladimir, and Ivan Petej. 2012. “Venn-Abers Predictors.” arXiv. https://doi.org/10.48550/ARXIV.1211.0025.
Vovk, Vladimir, Ivan Petej, Paolo Toccaceli, and Alex Gammerman. 2019. “Conformal Calibrators.” arXiv. https://doi.org/10.48550/ARXIV.1902.06579.
Vu, Tuan-Hung, Himalaya Jain, Maxime Bucher, Matthieu Cord, and Patrick Pérez. 2018. “ADVENT: Adversarial Entropy Minimization for Domain Adaptation in Semantic Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1811.12833.
Vyas, Nikhil, Sham Kakade, and Boaz Barak. 2023. “On Provable Copyright Protection for Generative Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.10870.
Wachinger, Christian, Martin Reuter, and Tassilo Klein. 2018. “DeepNAT: Deep Convolutional Neural Network for Segmenting Neuroanatomy.” NeuroImage 170 (April). https://doi.org/10.1016/j.neuroimage.2017.02.035.
Wager, Stefan, and Susan Athey. 2015. “Estimation and Inference of Heterogeneous Treatment Effects Using Random Forests.” arXiv. https://doi.org/10.48550/ARXIV.1510.04342.
Wake, Naoki, Atsushi Kanehira, Kazuhiro Sasabuchi, Jun Takamatsu, and Katsushi Ikeuchi. 2023. “ChatGPT Empowered Long-Step Robot Control in Various Environments: A Case Application.” IEEE Access 11. https://doi.org/10.1109/access.2023.3310935.
Walker, Jacob, Carl Doersch, Abhinav Gupta, and Martial Hebert. 2016. “An Uncertain Future: Forecasting from Static Images Using Variational Autoencoders.” arXiv. https://doi.org/10.48550/ARXIV.1606.07873.
Wallace, Eric, Shi Feng, Nikhil Kandpal, Matt Gardner, and Sameer Singh. 2019. “Universal Adversarial Triggers for Attacking and Analyzing NLP.” arXiv. https://doi.org/10.48550/ARXIV.1908.07125.
Wan, Fang, Pengxu Wei, Zhenjun Han, Jianbin Jiao, and Qixiang Ye. 2019. “Min-Entropy Latent Model for Weakly Supervised Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.1902.06057.
Wang, Andrew Z., Rex Ying, Pan Li, Nikhil Rao, Karthik Subbian, and Jure Leskovec. 2021. “Bipartite Dynamic Representations for Abuse Detection.” Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3447548.3467141.
Wang, Benjamin X., and Nathalie Japkowicz. 2009. “Boosting Support Vector Machines for Imbalanced Data Sets.” Knowledge and Information Systems 25 (March). https://doi.org/10.1007/s10115-009-0198-y.
Wang, Binghui, and Neil Zhenqiang Gong. 2019. “Attacking Graph-Based Classification via Manipulating the Graph Structure.” Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, November. https://doi.org/10.1145/3319535.3354206.
Wang, Boyu, and Joelle Pineau. 2013. “Online Ensemble Learning for Imbalanced Data Streams.” arXiv. https://doi.org/10.48550/ARXIV.1310.8004.
Wang, Bryan, Gang Li, and Yang Li. 2023. “Enabling Conversational Interaction with Mobile UI Using Large Language Models.” Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, April. https://doi.org/10.1145/3544548.3580895.
Wang, Chaofan, Samuel Kernan Freire, Mo Zhang, Jing Wei, Jorge Goncalves, Vassilis Kostakos, Zhanna Sarsenbayeva, Christina Schneegass, Alessandro Bozzon, and Evangelos Niforatos. 2023. “Safeguarding Crowdsourcing Surveys from ChatGPT with Prompt Injection.” arXiv. https://doi.org/10.48550/ARXIV.2306.08833.
Wang, Chaoqi, Roger Grosse, Sanja Fidler, and Guodong Zhang. 2019. “EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis.” arXiv. https://doi.org/10.48550/ARXIV.1905.05934.
Wang, Chengyi, Sanyuan Chen, Yu Wu, Ziqiang Zhang, Long Zhou, Shujie Liu, Zhuo Chen, et al. 2023. “Neural Codec Language Models Are Zero-Shot Text to Speech Synthesizers.” arXiv. https://doi.org/10.48550/ARXIV.2301.02111.
Wang, Chengyi, Shuangzhi Wu, and Shujie Liu. 2019. “Accelerating Transformer Decoding via a Hybrid of Self-Attention and Recurrent Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.1909.02279.
Wang, Chong, and David M. Blei. 2015. “A General Method for Robust Bayesian Modeling.” arXiv. https://doi.org/10.48550/ARXIV.1510.05078.
Wang, Daixin, Zhiqiang Zhang, Yeyu Zhao, Kai Huang, Yulin Kang, and Jun Zhou. 2023. “Financial Default Prediction via Motif-Preserving Graph Neural Network with Curriculum Learning.” Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3580305.3599351.
Wang, Danqing, Jiaze Chen, Xianze Wu, Hao Zhou, and Lei Li. 2021. “CNewSum: A Large-Scale Summarization Dataset with Human-Annotated Adequacy and Deducibility Level.” Natural Language Processing and Chinese Computing. https://doi.org/10.1007/978-3-030-88480-2_31.
Wang, Derek, Chaoran Li, Sheng Wen, Surya Nepal, and Yang Xiang. 2018. “Defending Against Adversarial Attack Towards Deep Neural Networks via Collaborative Multi-Task Training.” arXiv. https://doi.org/10.48550/ARXIV.1803.05123.
Wang, Haobo, Zhao Li, Jiaming Huang, Pengrui Hui, Weiwei Liu, Tianlei Hu, and Gang Chen. 2020. “Collaboration Based Multi-Label Propagation for Fraud Detection.” Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2020/343.
Wang, Hao, Tong Xu, Qi Liu, Defu Lian, Enhong Chen, Dongfang Du, Han Wu, and Wen Su. 2019. “MCNE.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330931.
Wang, Haoyu, Defu Lian, and Yong Ge. 2019. “Binarized Collaborative Filtering with Distilling Graph Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1906.01829.
Wang, Hongwei, Jia Wang, Jialin Wang, Miao Zhao, Weinan Zhang, Fuzheng Zhang, Xing Xie, and Minyi Guo. 2017. “GraphGAN: Graph Representation Learning with Generative Adversarial Nets.” arXiv. https://doi.org/10.48550/ARXIV.1711.08267.
Wang, Hongwei, Fuzheng Zhang, Min Hou, Xing Xie, Minyi Guo, and Qi Liu. 2018. “SHINE.” Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/3159652.3159666.
Wang, Hongwei, Fuzheng Zhang, Mengdi Zhang, Jure Leskovec, Miao Zhao, Wenjie Li, and Zhongyuan Wang. 2019. “Knowledge-Aware Graph Neural Networks with Label Smoothness Regularization for Recommender Systems.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330836.
Wang, Hongwei, Miao Zhao, Xing Xie, Wenjie Li, and Minyi Guo. 2019. “Knowledge Graph Convolutional Networks for Recommender Systems.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313417.
Wang, Huan, Qiming Zhang, Yuehai Wang, Yu Lu, and Haoji Hu. 2018. “Structured Pruning for Efficient ConvNets via Incremental Regularization.” arXiv. https://doi.org/10.48550/ARXIV.1804.09461.
Wang, Jingdong, Ting Zhang, Jingkuan Song, Nicu Sebe, and Heng Tao Shen. 2016. “A Survey on Learning to Hash.” arXiv. https://doi.org/10.48550/ARXIV.1606.00185.
Wang, Jingyuan, Ze Wang, Jianfeng Li, and Junjie Wu. 2018. “Multilevel Wavelet Decomposition Network for Interpretable Time Series Analysis.” arXiv. https://doi.org/10.48550/ARXIV.1806.08946.
WANG, Jixiang, Yunze LI, Xiangdong LIU, Chaoqun SHEN, Hongsheng ZHANG, and Kai XIONG. 2021. “Recent Active Thermal Management Technologies for the Development of Energy-Optimized Aerospace Vehicles in China.” Chinese Journal of Aeronautics 34 (February). https://doi.org/10.1016/j.cja.2020.06.021.
Wang, Jizhe, Pipei Huang, Huan Zhao, Zhibo Zhang, Binqiang Zhao, and Dik Lun Lee. 2018. “Billion-Scale Commodity Embedding for e-Commerce Recommendation in Alibaba.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3219869.
Wang, Jun, Benjamin Rubinstein, and Trevor Cohn. 2022. “Measuring and Mitigating Name Biases in Neural Machine Translation.” Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/2022.acl-long.184.
Wang, Jun, Lantao Yu, Weinan Zhang, Yu Gong, Yinghui Xu, Benyou Wang, Peng Zhang, and Dell Zhang. 2017. “IRGAN.” Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, August. https://doi.org/10.1145/3077136.3080786.
Wang, Jun, Lixing Zhu, Abhir Bhalerao, and Yulan He. 2023. “Can Prompt Learning Benefit Radiology Report Generation?” arXiv. https://doi.org/10.48550/ARXIV.2308.16269.
Wang, Ke, Senqiang Zhou, and Yu He. 2000. “Growing Decision Trees on Support-Less Association Rules.” Proceedings of the Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/347090.347147.
Wang, Kuan, Zhijian Liu, Yujun Lin, Ji Lin, and Song Han. 2018. “HAQ: Hardware-Aware Automated Quantization with Mixed Precision.” arXiv. https://doi.org/10.48550/ARXIV.1811.08886.
Wang, Lei, and Ee-Peng Lim. 2023. “Zero-Shot Next-Item Recommendation Using Large Pretrained Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.03153.
Wang, Lei, Chen Ma, Xueyang Feng, Zeyu Zhang, Hao Yang, Jingsen Zhang, Zhiyuan Chen, et al. 2023. “A Survey on Large Language Model Based Autonomous Agents.” arXiv. https://doi.org/10.48550/ARXIV.2308.11432.
Wang, Lewen, Haozhe Zhao, Cunguang Feng, Weiqing Liu, Congrui Huang, Marco Santoni, Manuel Cristofaro, Paola Jafrancesco, and Jiang Bian. 2023. “Removing Camouflage and Revealing Collusion: Leveraging Gang-Crime Pattern in Fraudster Detection.” Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3580305.3599895.
Wang, Limin, Yuanjun Xiong, Zhe Wang, Yu Qiao, Dahua Lin, Xiaoou Tang, and Luc Van Gool. 2017. “Temporal Segment Networks for Action Recognition in Videos.” arXiv. https://doi.org/10.48550/ARXIV.1705.02953.
Wang, Lin, Jake Elmstedt, Weng Kee Wong, and Hongquan Xu. 2021. “Orthogonal Subsampling for Big Data Linear Regression.” arXiv. https://doi.org/10.48550/ARXIV.2105.14647.
Wang, Luting, Yi Liu, Penghui Du, Zihan Ding, Yue Liao, Qiaosong Qi, Biaolong Chen, and Si Liu. 2023. “Object-Aware Distillation Pyramid for Open-Vocabulary Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2303.05892.
Wang, Lu, Zhengwu Zhang, and David Dunson. 2019. “Common and Individual Structure of Brain Networks.” The Annals of Applied Statistics 13 (March). https://doi.org/10.1214/18-aoas1193.
Wang, Min, Baoyuan Liu, and Hassan Foroosh. 2016. “Design of Efficient Convolutional Layers Using Single Intra-Channel Convolution, Topological Subdivisioning and Spatial "Bottleneck" Structure.” arXiv. https://doi.org/10.48550/ARXIV.1608.04337.
Wang, Mowei, Yong Cui, Xin Wang, Shihan Xiao, and Junchen Jiang. 2018. “Machine Learning for Networking: Workflow, Advances and Opportunities.” IEEE Network 32 (March). https://doi.org/10.1109/mnet.2017.1700200.
Wang, Panqu, Vicente Malave, and Ben Cipollini. 2015. “Encoding Voxels with Deep Learning.” The Journal of Neuroscience 35 (December). https://doi.org/10.1523/jneurosci.3454-15.2015.
Wang, Qixiang, Shanfeng Wang, Maoguo Gong, and Yue Wu. 2018. “Feature Hashing for Network Representation Learning.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/390.
Wang, Qizhou, Guansong Pang, Mahsa Salehi, Wray Buntine, and Christopher Leckie. 2023. “Open-Set Graph Anomaly Detection via Normal Structure Regularisation.” arXiv. https://doi.org/10.48550/ARXIV.2311.06835.
Wang, Quan, Pingping Huang, Haifeng Wang, Songtai Dai, Wenbin Jiang, Jing Liu, Yajuan Lyu, Yong Zhu, and Hua Wu. 2019. “CoKE: Contextualized Knowledge Graph Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1911.02168.
Wang, Rui, Hongsong Feng, and Guo-Wei Wei. 2023. “ChatGPT in Drug Discovery: A Case Study on Anticocaine Addiction Drug Development with Chatbots.” Journal of Chemical Information and Modeling 63 (November). https://doi.org/10.1021/acs.jcim.3c01429.
Wang, Ruoxi, Rakesh Shivanna, Derek Cheng, Sagar Jain, Dong Lin, Lichan Hong, and Ed Chi. 2021. “DCN V2: Improved Deep &Amp; Cross Network and Practical Lessons for Web-Scale Learning to Rank Systems.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3450078.
Wang, Sheng, Jian Peng, Jianzhu Ma, and Jinbo Xu. 2015. “Protein Secondary Structure Prediction Using Deep Convolutional Neural Fields.” arXiv. https://doi.org/10.48550/ARXIV.1512.00843.
Wang, Sheng, Siqi Sun, Zhen Li, Renyu Zhang, and Jinbo Xu. 2016. “Accurate de Novo Prediction of Protein Contact Map by Ultra-Deep Learning Model,” September. https://doi.org/10.1101/073239.
Wang, Sheng, Zihao Zhao, Xi Ouyang, Qian Wang, and Dinggang Shen. 2023. “ChatCAD: Interactive Computer-Aided Diagnosis on Medical Image Using Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.07257.
Wang, Shiqiang, Tiffany Tuor, Theodoros Salonidis, Kin K. Leung, Christian Makaya, Ting He, and Kevin Chan. 2018. “Adaptive Federated Learning in Resource Constrained Edge Computing Systems.” arXiv. https://doi.org/10.48550/ARXIV.1804.05271.
Wang, Shuhe, Xiaofei Sun, Xiaoya Li, Rongbin Ouyang, Fei Wu, Tianwei Zhang, Jiwei Li, and Guoyin Wang. 2023. “GPT-NER: Named Entity Recognition via Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2304.10428.
Wang, Shuohang, and Jing Jiang. 2016. “Machine Comprehension Using Match-LSTM and Answer Pointer.” arXiv. https://doi.org/10.48550/ARXIV.1608.07905.
Wang, Shuo, Zhe Li, Caiwen Ding, Bo Yuan, Qinru Qiu, Yanzhi Wang, and Yun Liang. 2018. “C-LSTM.” Proceedings of the 2018 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, February. https://doi.org/10.1145/3174243.3174253.
Wang, Suhang, Charu Aggarwal, Jiliang Tang, and Huan Liu. 2017. “Attributed Signed Network Embedding.” Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3132847.3132905.
Wang, Tianqi, Tong Geng, Ang Li, Xi Jin, and Martin Herbordt. 2020. “FPDeep: Scalable Acceleration of CNN Training on Deeply-Pipelined FPGA Clusters.” IEEE Transactions on Computers. https://doi.org/10.1109/tc.2020.3000118.
Wang, Tonghan, Tarun Gupta, Anuj Mahajan, Bei Peng, Shimon Whiteson, and Chongjie Zhang. 2020. “RODE: Learning Roles to Decompose Multi-Agent Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2010.01523.
Wang, Tonghan, Liang Zeng, Weijun Dong, Qianlan Yang, Yang Yu, and Chongjie Zhang. 2021. “Context-Aware Sparse Deep Coordination Graphs.” arXiv. https://doi.org/10.48550/ARXIV.2106.02886.
Wang, Wenguan, Xiankai Lu, Jianbing Shen, David Crandall, and Ling Shao. 2020. “Zero-Shot Video Object Segmentation via Attentive Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2001.06807.
Wang, Wenhai, Zhe Chen, Xiaokang Chen, Jiannan Wu, Xizhou Zhu, Gang Zeng, Ping Luo, et al. 2023. “VisionLLM: Large Language Model Is Also an Open-Ended Decoder for Vision-Centric Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2305.11175.
Wang, Wenhui, Hangbo Bao, Li Dong, Johan Bjorck, Zhiliang Peng, Qiang Liu, Kriti Aggarwal, et al. 2022. “Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2208.10442.
Wang, Wenjie, Minlie Huang, Xin-Shun Xu, Fumin Shen, and Liqiang Nie. 2018. “Chat More.” The 41st International ACM SIGIR Conference on Research &Amp; Development in Information Retrieval, June. https://doi.org/10.1145/3209978.3210061.
Wang, Wenjie, Xinyu Lin, Fuli Feng, Xiangnan He, and Tat-Seng Chua. 2023. “Generative Recommendation: Towards Next-Generation Recommender Paradigm.” arXiv. https://doi.org/10.48550/ARXIV.2304.03516.
Wang, Xiang, Xiangnan He, Yixin Cao, Meng Liu, and Tat-Seng Chua. 2019. “KGAT.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330989.
Wang, Xiao, Guangyao Chen, Guangwu Qian, Pengcheng Gao, Xiao-Yong Wei, Yaowei Wang, Yonghong Tian, and Wen Gao. 2023. “Large-Scale Multi-Modal Pre-Trained Models: A Comprehensive Survey.” Machine Intelligence Research 20 (June). https://doi.org/10.1007/s11633-022-1410-8.
Wang, Xiaolei, Kun Zhou, Ji-Rong Wen, and Wayne Xin Zhao. 2022. “Towards Unified Conversational Recommender Systems via Knowledge-Enhanced Prompt Learning.” Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3534678.3539382.
Wang, Xiaosong, Yifan Peng, Le Lu, Zhiyong Lu, and Ronald M. Summers. 2018. “TieNet: Text-Image Embedding Network for Common Thorax Disease Classification and Reporting in Chest x-Rays.” arXiv. https://doi.org/10.48550/ARXIV.1801.04334.
Wang, Xiaoyu, and Martin Benning. 2023. “Code Supporting [Lifted Bregman Training of Neural Networks],” August. https://doi.org/10.17863/CAM.86729.
Wang, Xiao, Meiqi Zhu, Deyu Bo, Peng Cui, Chuan Shi, and Jian Pei. 2020. “AM-GCN: Adaptive Multi-Channel Graph Convolutional Networks.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403177.
Wang, Xin, Geoffrey Oxholm, Da Zhang, and Yuan-Fang Wang. 2016. “Multimodal Transfer: A Hierarchical Deep Convolutional Neural Network for Fast Artistic Style Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1612.01895.
Wang, Xinxi, Yi Wang, David Hsu, and Ye Wang. 2013. “Exploration in Interactive Personalized Music Recommendation: A Reinforcement Learning Approach.” arXiv. https://doi.org/10.48550/ARXIV.1311.6355.
Wang, Xinyuan, Chenxi Li, Zhen Wang, Fan Bai, Haotian Luo, Jiayou Zhang, Nebojsa Jojic, Eric P. Xing, and Zhiting Hu. 2023. “PromptAgent: Strategic Planning with Language Models Enables Expert-Level Prompt Optimization.” arXiv. https://doi.org/10.48550/ARXIV.2310.16427.
Wang, Xuhong, Ding Lyu, Mengjian Li, Yang Xia, Qi Yang, Xinwen Wang, Xinguang Wang, et al. 2021. “APAN: Asynchronous Propagation Attention Network for Real-Time Temporal Graph Embedding.” Proceedings of the 2021 International Conference on Management of Data, June. https://doi.org/10.1145/3448016.3457564.
Wang, Yanshan, Sijia Liu, Naveed Afzal, Majid Rastegar-Mojarad, Liwei Wang, Feichen Shen, Paul Kingsbury, and Hongfang Liu. 2018. “A Comparison of Word Embeddings for the Biomedical Natural Language Processing.” arXiv. https://doi.org/10.48550/ARXIV.1802.00400.
Wang, Yaqing, Quanming Yao, James Kwok, and Lionel M. Ni. 2019. “Generalizing from a Few Examples: A Survey on Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.1904.05046.
Wang, Yen-Jen, Bike Zhang, Jianyu Chen, and Koushil Sreenath. 2023. “Prompt a Robot to Walk with Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2309.09969.
Wang, Yida, David Joseph Tan, Nassir Navab, and Federico Tombari. 2022. “SoftPool++: An Encoder–Decoder Network for Point Cloud Completion.” International Journal of Computer Vision 130 (March). https://doi.org/10.1007/s11263-022-01588-7.
Wang, Yubin, Xinyang Jiang, De Cheng, Dongsheng Li, and Cairong Zhao. 2023. “Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2312.06323.
Wang, Yue, Yongbin Sun, Ziwei Liu, Sanjay E. Sarma, Michael M. Bronstein, and Justin M. Solomon. 2019. “Dynamic Graph CNN for Learning on Point Clouds.” ACM Transactions on Graphics 38 (October). https://doi.org/10.1145/3326362.
Wang, Yue, Dawei Yin, Luo Jie, Pengyuan Wang, Makoto Yamada, Yi Chang, and Qiaozhu Mei. 2016. “Beyond Ranking.” Proceedings of the Ninth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/2835776.2835824.
Wang, Yu, Jiayi Liu, Yuxiang Liu, Jun Hao, Yang He, Jinghe Hu, Weipeng P. Yan, and Mantian Li. 2017. “LADDER: A Human-Level Bidding Agent for Large-Scale Real-Time Online Auctions.” arXiv. https://doi.org/10.48550/ARXIV.1708.05565.
Wang, Yuqi, Yuntao Chen, and Zhaoxiang Zhang. 2022. “4D Unsupervised Object Discovery.” arXiv. https://doi.org/10.48550/ARXIV.2210.04801.
Wang, Yuxuan, RJ Skerry-Ryan, Ying Xiao, Daisy Stanton, Joel Shor, Eric Battenberg, Rob Clark, and Rif A. Saurous. 2017. “Uncovering Latent Style Factors for Expressive Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1711.00520.
Wang, Yuxuan, Daisy Stanton, Yu Zhang, RJ Skerry-Ryan, Eric Battenberg, Joel Shor, Ying Xiao, Fei Ren, Ye Jia, and Rif A. Saurous. 2018. “Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1803.09017.
Wang, Zhen, Rameswar Panda, Leonid Karlinsky, Rogerio Feris, Huan Sun, and Yoon Kim. 2023. “Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning.” arXiv. https://doi.org/10.48550/ARXIV.2303.02861.
Wang, Zhe, Liqin Zhao, Biye Jiang, Guorui Zhou, Xiaoqiang Zhu, and Kun Gai. 2020. “COLD: Towards the Next Generation of Pre-Ranking System.” arXiv. https://doi.org/10.48550/ARXIV.2007.16122.
Wang, Zhiguang, Weizhong Yan, and Tim Oates. 2016. “Time Series Classification from Scratch with Deep Neural Networks: A Strong Baseline.” arXiv. https://doi.org/10.48550/ARXIV.1611.06455.
Wang, Zhiguo, Haitao Mi, Wael Hamza, and Radu Florian. 2016. “Multi-Perspective Context Matching for Machine Comprehension.” arXiv. https://doi.org/10.48550/ARXIV.1612.04211.
Wang, Zhongdao, Liang Zheng, Yali Li, and Shengjin Wang. 2019. “Linkage Based Face Clustering via Graph Convolution Network.” arXiv. https://doi.org/10.48550/ARXIV.1903.11306.
Wang, Zi, Chengcheng Li, Dali Wang, Xiangyang Wang, and Hairong Qi. 2019. “Speeding up Convolutional Networks Pruning with Coarse Ranking.” arXiv. https://doi.org/10.48550/ARXIV.1902.06385.
Wang, Ziyu, Tom Schaul, Matteo Hessel, Hado van Hasselt, Marc Lanctot, and Nando de Freitas. 2015. “Dueling Network Architectures for Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1511.06581.
Wangperawong, Artit, Cyrille Brun, Olav Laudy, and Rujikorn Pavasuthipaisit. 2016. “Churn Analysis Using Deep Convolutional Neural Networks and Autoencoders.” arXiv. https://doi.org/10.48550/ARXIV.1604.05377.
Wardat, Yousef, Mohammad A. Tashtoush, Rommel AlAli, and Adeeb M. Jarrah. 2023. “ChatGPT: A Revolutionary Tool for Teaching and Learning Mathematics.” Eurasia Journal of Mathematics, Science and Technology Education 19 (July). https://doi.org/10.29333/ejmste/13272.
Warden, Pete, Matthew Stewart, Brian Plancher, Colby Banbury, Shvetank Prakash, Emma Chen, Zain Asgar, Sachin Katti, and Vijay Janapa Reddi. 2022. “Machine Learning Sensors.” arXiv. https://doi.org/10.48550/ARXIV.2206.03266.
Wastl, Michelle, Jannis Vamvas, and Rico Sennrich. 2024. “Machine Translation Models Are Zero-Shot Detectors of Translation Direction.” arXiv. https://doi.org/10.48550/ARXIV.2401.06769.
Watcharapichat, Pijika, Victoria Lopez Morales, Raul Castro Fernandez, and Peter Pietzuch. 2016. “Ako.” Proceedings of the Seventh ACM Symposium on Cloud Computing, October. https://doi.org/10.1145/2987550.2987586.
Watson, Andrew B., and Denis G. Pelli. 1983. “Quest: A Bayesian Adaptive Psychometric Method.” Perception &Amp; Psychophysics 33 (March). https://doi.org/10.3758/bf03202828.
Watson, Daniel, William Chan, Ricardo Martin-Brualla, Jonathan Ho, Andrea Tagliasacchi, and Mohammad Norouzi. 2022. “Novel View Synthesis with Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2210.04628.
Webb, Taylor, Keith J. Holyoak, and Hongjing Lu. 2022. “Emergent Analogical Reasoning in Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2212.09196.
Weber, Mark, Giacomo Domeniconi, Jie Chen, Daniel Karl I. Weidele, Claudio Bellei, Tom Robinson, and Charles E. Leiserson. 2019. “Anti-Money Laundering in Bitcoin: Experimenting with Graph Convolutional Networks for Financial Forensics.” arXiv. https://doi.org/10.48550/ARXIV.1908.02591.
Wei, Alexander, Nika Haghtalab, and Jacob Steinhardt. 2023. “Jailbroken: How Does LLM Safety Training Fail?” arXiv. https://doi.org/10.48550/ARXIV.2307.02483.
Wei, Duan, and Li Lin. 2019. “An External Knowledge Enhanced Multi-Label Charge Prediction Approach with Label Number Learning.” arXiv. https://doi.org/10.48550/ARXIV.1907.02205.
Wei, Jerry, Jason Wei, Yi Tay, Dustin Tran, Albert Webson, Yifeng Lu, Xinyun Chen, et al. 2023. “Larger Language Models Do in-Context Learning Differently.” arXiv. https://doi.org/10.48550/ARXIV.2303.03846.
Wei, Kang, Jun Li, Ming Ding, Chuan Ma, Howard H. Yang, Farokhi Farhad, Shi Jin, Tony Q. S. Quek, and H. Vincent Poor. 2019. “Federated Learning with Differential Privacy: Algorithms and Performance Analysis.” arXiv. https://doi.org/10.48550/ARXIV.1911.00222.
Wei, Qi, Nicolas Dobigeon, and Jean-Yves Tourneret. 2013. “Bayesian Fusion of Multi-Band Images.” arXiv. https://doi.org/10.48550/ARXIV.1307.5996.
Wei, Ran, Siliang Zeng, Chenliang Li, Alfredo Garcia, Anthony McDonald, and Mingyi Hong. 2023. “A Bayesian Approach to Robust Inverse Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.2309.08571.
Wei, Xiang, Xingyu Cui, Ning Cheng, Xiaobin Wang, Xin Zhang, Shen Huang, Pengjun Xie, et al. 2023. “Zero-Shot Information Extraction via Chatting with ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2302.10205.
Wei, Xuechao, Cody Hao Yu, Peng Zhang, Youxiang Chen, Yuxin Wang, Han Hu, Yun Liang, and Jason Cong. 2017. “Automated Systolic Array Architecture Synthesis for High Throughput CNN Inference on FPGAs.” Proceedings of the 54th Annual Design Automation Conference 2017, June. https://doi.org/10.1145/3061639.3062207.
Wei, Xue-Xin, and Alan A Stocker. 2015. “A Bayesian Observer Model Constrained by Efficient Coding Can Explain ’Anti-Bayesian’ Percepts.” Nature Neuroscience 18 (September). https://doi.org/10.1038/nn.4105.
Weiler, Maurice, Fred A. Hamprecht, and Martin Storath. 2017. “Learning Steerable Filters for Rotation Equivariant CNNs.” arXiv. https://doi.org/10.48550/ARXIV.1711.07289.
Weinberger, Kilian, Anirban Dasgupta, Josh Attenberg, John Langford, and Alex Smola. 2009. “Feature Hashing for Large Scale Multitask Learning.” arXiv. https://doi.org/10.48550/ARXIV.0902.2206.
Weisz, Justin D., Michael Muller, Jessica He, and Stephanie Houde. 2023. “Toward General Design Principles for Generative AI Applications.” arXiv. https://doi.org/10.48550/ARXIV.2301.05578.
Wen, Liangjian, Xuanyang Zhang, Haoli Bai, and Zenglin Xu. 2019. “Structured Pruning of Recurrent Neural Networks Through Neuron Selection.” arXiv. https://doi.org/10.48550/ARXIV.1906.06847.
Wen, Tsung-Hsien, Milica Gasic, Nikola Mrksic, Pei-Hao Su, David Vandyke, and Steve Young. 2015. “Semantically Conditioned LSTM-Based Natural Language Generation for Spoken Dialogue Systems.” arXiv. https://doi.org/10.48550/ARXIV.1508.01745.
Wen, Zhihao, and Yuan Fang. 2023. “Prompt Tuning on Graph-Augmented Low-Resource Text Classification.” arXiv. https://doi.org/10.48550/ARXIV.2307.10230.
Wen, Zhihao, Yuan Fang, Yihan Liu, Yang Guo, and Shuji Hao. 2023. “Voucher Abuse Detection with Prompt-Based Fine-Tuning on Graph Neural Networks.” Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3583780.3615505.
Weng, Haiqin, Zhao Li, Shouling Ji, Chen Chu, Haifeng Lu, Tianyu Du, and Qinming He. 2018. “Online e-Commerce Fraud: A Large-Scale Detection and Analysis.” 2018 IEEE 34th International Conference on Data Engineering (ICDE), April. https://doi.org/10.1109/icde.2018.00162.
Weng, Thomas, David Held, Franziska Meier, and Mustafa Mukadam. 2022. “Neural Grasp Distance Fields for Robot Manipulation.” arXiv. https://doi.org/10.48550/ARXIV.2211.02647.
West, Mike, P. Jeff Harrison, and Helio S. Migon. 1985. “Dynamic Generalized Linear Models and Bayesian Forecasting.” Journal of the American Statistical Association 80 (March). https://doi.org/10.1080/01621459.1985.10477131.
Westgate, Bradford S., Dawn B. Woodard, David S. Matteson, and Shane G. Henderson. 2013. “Travel Time Estimation for Ambulances Using Bayesian Data Augmentation.” The Annals of Applied Statistics 7 (June). https://doi.org/10.1214/13-aoas626.
Weston, Jason, Antoine Bordes, Sumit Chopra, Alexander M. Rush, Bart van Merriënboer, Armand Joulin, and Tomas Mikolov. 2015. “Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks.” arXiv. https://doi.org/10.48550/ARXIV.1502.05698.
Weston, Jason, Sumit Chopra, and Antoine Bordes. 2014. “Memory Networks.” arXiv. https://doi.org/10.48550/ARXIV.1410.3916.
Weston, Jason, Hector Yee, and Ron J. Weiss. 2013. “Learning to Rank Recommendations with the k-Order Statistic Loss.” Proceedings of the 7th ACM Conference on Recommender Systems, October. https://doi.org/10.1145/2507157.2507210.
White, Andrew D., Glen M. Hocky, Heta A. Gandhi, Mehrad Ansari, Sam Cox, Geemi P. Wellawatte, Subarna Sasmal, et al. 2022a. “Do Large Language Models Know Chemistry?” July. https://doi.org/10.26434/chemrxiv-2022-3md3n.
———, et al. 2022b. “Assessment of Chemistry Knowledge in Large Language Models That Generate Code,” December. https://doi.org/10.26434/chemrxiv-2022-3md3n-v2.
White, Jules, Quchen Fu, Sam Hays, Michael Sandborn, Carlos Olea, Henry Gilbert, Ashraf Elnashar, Jesse Spencer-Smith, and Douglas C. Schmidt. 2023. “A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2302.11382.
White, Jules, Sam Hays, Quchen Fu, Jesse Spencer-Smith, and Douglas C. Schmidt. 2023. “ChatGPT Prompt Patterns for Improving Code Quality, Refactoring, Requirements Elicitation, and Software Design.” arXiv. https://doi.org/10.48550/ARXIV.2303.07839.
Wiatowski, Thomas, and Helmut Bölcskei. 2015. “A Mathematical Theory of Deep Convolutional Neural Networks for Feature Extraction.” arXiv. https://doi.org/10.48550/ARXIV.1512.06293.
Wick, Christoph, Jochen Zöllner, and Tobias Grüning. 2021. “Rescoring Sequence-to-Sequence Models for Text Line Recognition with CTC-Prefixes.” arXiv. https://doi.org/10.48550/ARXIV.2110.05909.
Williams, Jason D., and Geoffrey Zweig. 2016. “End-to-End LSTM-Based Dialog Control Optimized with Supervised and Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.01269.
Williams, Melinda Han, Claudia Perlich, Brian Dalessandro, and Foster Provost. 2014. “Pleasing the Advertising Oracle.” Proceedings of the Eighth International Workshop on Data Mining for Online Advertising, August. https://doi.org/10.1145/2648584.2648587.
Williams, Samuel, Andrew Waterman, and David Patterson. 2009. “Roofline.” Communications of the ACM 52 (April). https://doi.org/10.1145/1498765.1498785.
Williamson, Ben, Felicitas Macgilchrist, and John Potter. 2023. “Re-Examining AI, Automation and Datafication in Education.” Learning, Media and Technology 48 (January). https://doi.org/10.1080/17439884.2023.2167830.
Wilson, Andrew Gordon, Zhiting Hu, Ruslan Salakhutdinov, and Eric P. Xing. 2016. “Stochastic Variational Deep Kernel Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.00336.
Wingate, David, Mohammad Shoeybi, and Taylor Sorensen. 2022. “Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2210.03162.
Wiseman, Sam, and Alexander M. Rush. 2016. “Sequence-to-Sequence Learning as Beam-Search Optimization.” arXiv. https://doi.org/10.48550/ARXIV.1606.02960.
Wiseman, Sam, Stuart M. Shieber, and Alexander M. Rush. 2017. “Challenges in Data-to-Document Generation.” arXiv. https://doi.org/10.48550/ARXIV.1707.08052.
Wistuba, Martin, Nicolas Schilling, and Lars Schmidt-Thieme. 2017. “Automatic Frankensteining: Creating Complex Ensembles Autonomously.” Proceedings of the 2017 SIAM International Conference on Data Mining, June. https://doi.org/10.1137/1.9781611974973.83.
Witteveen, Sam, and Martin Andrews. 2022. “Investigating Prompt Engineering in Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2211.15462.
Woellner, Cristina, E. Michael Gertz, Alejandro A. Schäffer, Macarena Lagos, Mario Perro, Erik-Oliver Glocker, Maria C. Pietrogrande, et al. 2010. “Mutations in STAT3 and Diagnostic Guidelines for Hyper-IgE Syndrome.” Journal of Allergy and Clinical Immunology 125 (February). https://doi.org/10.1016/j.jaci.2009.10.059.
Wolf, Thomas, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, et al. 2019. “HuggingFace’s Transformers: State-of-the-Art Natural Language Processing.” arXiv. https://doi.org/10.48550/ARXIV.1910.03771.
Wolpert, Daniel M., Zoubin Ghahramani, and Michael I. Jordan. 1995. “An Internal Model for Sensorimotor Integration.” Science 269 (September). https://doi.org/10.1126/science.7569931.
Wolpert, David H. 2023. “The Implications of the No-Free-Lunch Theorems for Meta-Induction.” Journal for General Philosophy of Science 54 (March). https://doi.org/10.1007/s10838-022-09609-2.
Wondimu, Natnael A., Cédric Buche, and Ubbo Visser. 2022. “Interactive Machine Learning: A State of the Art Review.” arXiv. https://doi.org/10.48550/ARXIV.2207.06196.
Wong, Jeffrey C. 2020. “Computational Causal Inference.” arXiv. https://doi.org/10.48550/ARXIV.2007.10979.
Wong, Sebastien C., Adam Gatt, Victor Stamatescu, and Mark D. McDonnell. 2016. “Understanding Data Augmentation for Classification: When to Warp?” arXiv. https://doi.org/10.48550/ARXIV.1609.08764.
Workshop, BigScience, Teven Le Scao, Angela Fan, Christopher Akiki, Ellie Pavlick, Suzana Ilić, Daniel Hesslow, et al. 2022. “BLOOM: A 176B-Parameter Open-Access Multilingual Language Model.” arXiv. https://doi.org/10.48550/ARXIV.2211.05100.
Wright, Marvin N., and Andreas Ziegler. 2017. “<B>ranger</b>: A Fast Implementation of Random Forests for High Dimensional Data in <i>c++</i> and <i>r</i>.” Journal of Statistical Software 77. https://doi.org/10.18637/jss.v077.i01.
Wu, Cheng-En, Yu Tian, Haichao Yu, Heng Wang, Pedro Morgado, Yu Hen Hu, and Linjie Yang. 2023. “Why Is Prompt Tuning for Vision-Language Models Robust to Noisy Labels?” arXiv. https://doi.org/10.48550/ARXIV.2307.11978.
Wu, Chenxi, Min Zhu, Qinyang Tan, Yadhu Kartha, and Lu Lu. 2023. “A Comprehensive Study of Non-Adaptive and Residual-Based Adaptive Sampling for Physics-Informed Neural Networks.” Computer Methods in Applied Mechanics and Engineering 403 (January). https://doi.org/10.1016/j.cma.2022.115671.
Wu, Chien-Sheng, Steven Hoi, Richard Socher, and Caiming Xiong. 2020. “TOD-BERT: Pre-Trained Natural Language Understanding for Task-Oriented Dialogue.” arXiv. https://doi.org/10.48550/ARXIV.2004.06871.
Wu, Di, Xiujun Chen, Xun Yang, Hao Wang, Qing Tan, Xiaoxun Zhang, Jian Xu, and Kun Gai. 2018. “Budget Constrained Bidding by Model-Free Reinforcement Learning in Display Advertising.” Proceedings of the 27th ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3269206.3271748.
Wu, Felix, Kwangyoun Kim, Shinji Watanabe, Kyu Han, Ryan McDonald, Kilian Q. Weinberger, and Yoav Artzi. 2022. “Wav2Seq: Pre-Training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages.” arXiv. https://doi.org/10.48550/ARXIV.2205.01086.
Wu, Hui, Michele Merler, Rosario Uceda-Sosa, and John R. Smith. 2016. “Learning to Make Better Mistakes.” Proceedings of the 24th ACM International Conference on Multimedia, October. https://doi.org/10.1145/2964284.2967205.
Wu, Jay Zhangjie, Xiuyu Li, Difei Gao, Zhen Dong, Jinbin Bai, Aishani Singh, Xiaoyu Xiang, et al. 2023. “CVPR 2023 Text Guided Video Editing Competition.” arXiv. https://doi.org/10.48550/ARXIV.2310.16003.
Wu, Jiaxiang, Cong Leng, Yuhang Wang, Qinghao Hu, and Jian Cheng. 2015. “Quantized Convolutional Neural Networks for Mobile Devices.” arXiv. https://doi.org/10.48550/ARXIV.1512.06473.
Wu, Jiayang, Wensheng Gan, Zefeng Chen, Shicheng Wan, and Hong Lin. 2023. “AI-Generated Content (AIGC): A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2304.06632.
Wu, Junfei, Weizhi Xu, Qiang Liu, Shu Wu, and Liang Wang. 2022. “Adversarial Contrastive Learning for Evidence-Aware Fake News Detection with Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2210.05498.
Wu, Jun, Jingrui He, and Jiejun Xu. 2019. “DEMO-Net: Degree-Specific Graph Neural Networks for Node and Graph Classification.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330950.
Wu, Jyun-Yi, Cheng Yu, Szu-Wei Fu, Chih-Ting Liu, Shao-Yi Chien, and Yu Tsao. 2019. “Increasing Compactness of Deep Learning Based Speech Enhancement Models with Parameter Pruning and Quantization Techniques.” IEEE Signal Processing Letters 26 (December). https://doi.org/10.1109/lsp.2019.2951950.
Wu, Longfeng, Bowen Lei, Dongkuan Xu, and Dawei Zhou. 2023. “Towards Reliable Rare Category Analysis on Graphs via Individual Calibration.” Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3580305.3599525.
Wu, Mingrui, Xuying Zhang, Xiaoshuai Sun, Yiyi Zhou, Chao Chen, Jiaxin Gu, Xing Sun, and Rongrong Ji. 2022. “DIFNet: Boosting Visual Information Flow for Image Captioning.” 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52688.2022.01749.
Wu, Qitian, Hengrui Zhang, Xiaofeng Gao, Peng He, Paul Weng, Han Gao, and Guihai Chen. 2019. “Dual Graph Attention Networks for Deep Latent Representation of Multifaceted Social Effects in Recommender Systems.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313442.
Wu, Rundi, and Changxi Zheng. 2022. “Learning to Generate 3D Shapes from a Single Example.” ACM Transactions on Graphics 41 (November). https://doi.org/10.1145/3550454.3555480.
Wu, Shuang, Guoqi Li, Feng Chen, and Luping Shi. 2018. “Training and Inference with Integers in Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1802.04680.
Wu, Tao, Ellie Ka-In Chio, Heng-Tze Cheng, Yu Du, Steffen Rendle, Dima Kuzmin, Ritesh Agarwal, et al. 2020. “Zero-Shot Heterogeneous Transfer Learning from Recommender Systems to Cold-Start Search Retrieval.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3412752.
Wu, Wenquan, Zhen Guo, Xiangyang Zhou, Hua Wu, Xiyuan Zhang, Rongzhong Lian, and Haifeng Wang. 2019. “Proactive Human-Machine Conversation with Explicit Conversation Goals.” arXiv. https://doi.org/10.48550/ARXIV.1906.05572.
Wu, Wentao, Xiaoxu Zhu, Jiaming Tao, and Peifeng Li. 2018. “Event Detection via Recurrent Neural Network and Argument Prediction.” Natural Language Processing and Chinese Computing. https://doi.org/10.1007/978-3-319-99501-4_20.
Wu, Xian, Stephen Mattingly, Shayan Mirjafari, Chao Huang, and Nitesh V. Chawla. 2020. “Personalized Imputation on Wearable-Sensory Time Series via Knowledge Transfer.” Proceedings of the 29th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3340531.3411879.
Wu, Xindong, Vipin Kumar, J. Ross Quinlan, Joydeep Ghosh, Qiang Yang, Hiroshi Motoda, Geoffrey J. McLachlan, et al. 2007. “Top 10 Algorithms in Data Mining.” Knowledge and Information Systems 14 (December). https://doi.org/10.1007/s10115-007-0114-2.
Wu, Xinle, Dalin Zhang, Chenjuan Guo, Chaoyang He, Bin Yang, and Christian S. Jensen. 2021. “AutoCTS.” Proceedings of the VLDB Endowment 15 (December). https://doi.org/10.14778/3503585.3503604.
Wu, Yiqing, Ruobing Xie, Yongchun Zhu, Fuzhen Zhuang, Xu Zhang, Leyu Lin, and Qing He. 2022. “Personalized Prompt for Sequential Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.2205.09666.
Wu, Yonghui, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, et al. 2016. “Google’s Neural Machine Translation System: Bridging the Gap Between Human and Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1609.08144.
Wu, Yongji, Defu Lian, Yiheng Xu, Le Wu, and Enhong Chen. 2020. “Graph Convolutional Networks with Markov Random Field Reasoning for Social Spammer Detection.” Proceedings of the AAAI Conference on Artificial Intelligence 34 (April). https://doi.org/10.1609/aaai.v34i01.5455.
Wu, Yusong, Ke Chen, Tianyu Zhang, Yuchen Hui, Taylor Berg-Kirkpatrick, and Shlomo Dubnov. 2022. “Large-Scale Contrastive Language-Audio Pretraining with Feature Fusion and Keyword-to-Caption Augmentation.” arXiv. https://doi.org/10.48550/ARXIV.2211.06687.
Wu, Yuxin, and Justin Johnson. 2021. “Rethinking "Batch" in BatchNorm.” arXiv. https://doi.org/10.48550/ARXIV.2105.07576.
Wu, Zhenhe, Xiaoguang Yu, Meng Chen, Liangqing Wu, Jiahao Ji, and Zhoujun Li. 2023. “Enhancing New Intent Discovery via Robust Neighbor-Based Contrastive Learning.” INTERSPEECH 2023, August. https://doi.org/10.21437/interspeech.2023-1740.
Wu, Zhenqin, Bharath Ramsundar, Evan N. Feinberg, Joseph Gomes, Caleb Geniesse, Aneesh S. Pappu, Karl Leswing, and Vijay Pande. 2017. “MoleculeNet: A Benchmark for Molecular Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1703.00564.
Wu, Zhijin, Rafael A Irizarry, Robert Gentleman, Francisco Martinez-Murillo, and Forrest Spencer. 2004. “A Model-Based Background Adjustment for Oligonucleotide Expression Arrays.” Journal of the American Statistical Association 99 (December). https://doi.org/10.1198/016214504000000683.
Wu, Zihao, Lu Zhang, Chao Cao, Xiaowei Yu, Haixing Dai, Chong Ma, Zhengliang Liu, et al. 2023. “Exploring the Trade-Offs: Unified Large Language Models Vs Local Fine-Tuned Models for Highly-Specific Radiology NLI Task.” arXiv. https://doi.org/10.48550/ARXIV.2304.09138.
Xia, Chunqiu Steven, and Lingming Zhang. 2023. “Conversational Automated Program Repair.” arXiv. https://doi.org/10.48550/ARXIV.2301.13246.
Xia, Yingce, Di He, Tao Qin, Liwei Wang, Nenghai Yu, Tie-Yan Liu, and Wei-Ying Ma. 2016. “Dual Learning for Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1611.00179.
Xiang, Chong, Charles R. Qi, and Bo Li. 2018. “Generating 3D Adversarial Point Clouds.” arXiv. https://doi.org/10.48550/ARXIV.1809.07016.
Xiao, Aoran, Jiaxing Huang, Dayan Guan, Xiaoqin Zhang, Shijian Lu, and Ling Shao. 2023. “Unsupervised Point Cloud Representation Learning with Deep Neural Networks: A Survey.” IEEE Transactions on Pattern Analysis and Machine Intelligence 45 (September). https://doi.org/10.1109/tpami.2023.3262786.
Xiao, Chaowei, Bo Li, Jun-Yan Zhu, Warren He, Mingyan Liu, and Dawn Song. 2018. “Generating Adversarial Examples with Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1801.02610.
Xiao, Guangxuan, Yuandong Tian, Beidi Chen, Song Han, and Mike Lewis. 2023. “Efficient Streaming Language Models with Attention Sinks.” arXiv. https://doi.org/10.48550/ARXIV.2309.17453.
Xiao, Han, Minlie Huang, Yu Hao, and Xiaoyan Zhu. 2015. “TransA: An Adaptive Approach for Knowledge Graph Embedding.” arXiv. https://doi.org/10.48550/ARXIV.1509.05490.
Xiao, Han, Minlie Huang, and Xiaoyan Zhu. 2015. “From One Point to a Manifold: Knowledge Graph Embedding for Precise Link Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1512.04792.
Xiao, Huang, Battista Biggio, Gavin Brown, Giorgio Fumera, Claudia Eckert, and Fabio Roli. 2018. “Is Feature Selection Secure Against Training Data Poisoning?” arXiv. https://doi.org/10.48550/ARXIV.1804.07933.
Xiao, Qingcheng, Yun Liang, Liqiang Lu, Shengen Yan, and Yu-Wing Tai. 2017. “Exploring Heterogeneous Algorithms for Accelerating Deep Convolutional Neural Networks on FPGAs.” Proceedings of the 54th Annual Design Automation Conference 2017, June. https://doi.org/10.1145/3061639.3062244.
Xiao, Yao, Ziyi Tang, Pengxu Wei, Cong Liu, and Liang Lin. 2023. “Masked Images Are Counterfactual Samples for Robust Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2303.03052.
Xiao, Yisheng, Lijun Wu, Junliang Guo, Juntao Li, Min Zhang, Tao Qin, and Tie-yan Liu. 2022. “A Survey on Non-Autoregressive Generation for Neural Machine Translation and Beyond.” arXiv. https://doi.org/10.48550/ARXIV.2204.09269.
Xie, Cihang, Jianyu Wang, Zhishuai Zhang, Zhou Ren, and Alan Yuille. 2017. “Mitigating Adversarial Effects Through Randomization.” arXiv. https://doi.org/10.48550/ARXIV.1711.01991.
Xie, Fangzheng, Joshua Cape, Carey E. Priebe, and Yanxun Xu. 2022. “Bayesian Sparse Spiked Covariance Model with a Continuous Matrix Shrinkage Prior.” Bayesian Analysis 17 (December). https://doi.org/10.1214/21-ba1292.
Xie, Saining, Ross Girshick, Piotr Dollár, Zhuowen Tu, and Kaiming He. 2016. “Aggregated Residual Transformations for Deep Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1611.05431.
Xie, Sang Michael, Aditi Raghunathan, Percy Liang, and Tengyu Ma. 2021. “An Explanation of in-Context Learning as Implicit Bayesian Inference.” arXiv. https://doi.org/10.48550/ARXIV.2111.02080.
Xie, Xijiong, and Shiliang Sun. 2015. “Multi-View Twin Support Vector Machines.” Intelligent Data Analysis 19 (July). https://doi.org/10.3233/ida-150740.
Xie, Yaqi, Chen Yu, Tongyao Zhu, Jinbin Bai, Ze Gong, and Harold Soh. 2023. “Translating Natural Language to Planning Goals with Large-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2302.05128.
Xie, Zhiwen, Runjie Zhu, Kunsong Zhao, Jin Liu, Guangyou Zhou, and Jimmy Xiangji Huang. 2020. “A Contextual Alignment Enhanced Cross Graph Attention Network for Cross-Lingual Entity Alignment.” Proceedings of the 28th International Conference on Computational Linguistics. https://doi.org/10.18653/v1/2020.coling-main.520.
Xin, Kexuan, Zequn Sun, Wen Hua, Wei Hu, and Xiaofang Zhou. 2022. “Informed Multi-Context Entity Alignment.” Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/3488560.3498523.
Xing, Eric P., Qirong Ho, Wei Dai, Jin Kyu Kim, Jinliang Wei, Seunghak Lee, Xun Zheng, Pengtao Xie, Abhimanu Kumar, and Yaoliang Yu. 2013. “Petuum: A New Platform for Distributed Machine Learning on Big Data,” December. http://arxiv.org/abs/1312.7651v2.
Xiong, Caiming, Victor Zhong, and Richard Socher. 2016. “Dynamic Coattention Networks for Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1611.01604.
Xiong, Honglin, Sheng Wang, Yitao Zhu, Zihao Zhao, Yuxiao Liu, Linlin Huang, Qian Wang, and Dinggang Shen. 2023. “DoctorGLM: Fine-Tuning Your Chinese Doctor Is Not a Herculean Task.” arXiv. https://doi.org/10.48550/ARXIV.2304.01097.
Xiong, W., J. Droppo, X. Huang, F. Seide, M. Seltzer, A. Stolcke, D. Yu, and G. Zweig. 2016. “Achieving Human Parity in Conversational Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1610.05256.
Xu, Canwen, Daya Guo, Nan Duan, and Julian McAuley. 2023. “Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data.” arXiv. https://doi.org/10.48550/ARXIV.2304.01196.
Xu, Chao, Zhentan Feng, Yizheng Chen, Minghua Wang, and Tao Wei. 2018. “FeatNet.” Proceedings of the 11th ACM Workshop on Artificial Intelligence and Security, January. https://doi.org/10.1145/3270101.3270109.
Xu, Chen, Jianqiang Yao, Zhouchen Lin, Wenwu Ou, Yuanbin Cao, Zhirong Wang, and Hongbin Zha. 2018. “Alternating Multi-Bit Quantization for Recurrent Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1802.00150.
Xu, Fan, Nan Wang, Xuezhi Wen, Meiqi Gao, Chaoqun Guo, and Xibin Zhao. 2023. “Few-Shot Message-Enhanced Contrastive Learning for Graph Anomaly Detection.” arXiv. https://doi.org/10.48550/ARXIV.2311.10370.
Xu, Han, Yao Ma, Hao-Chen Liu, Debayan Deb, Hui Liu, Ji-Liang Tang, and Anil K. Jain. 2020. “Adversarial Attacks and Defenses in Images, Graphs and Text: A Review.” International Journal of Automation and Computing 17 (March). https://doi.org/10.1007/s11633-019-1211-x.
Xu, Haowen, Yang Feng, Jie Chen, Zhaogang Wang, Honglin Qiao, Wenxiao Chen, Nengwen Zhao, et al. 2018. “Unsupervised Anomaly Detection via Variational Auto-Encoder for Seasonal KPIs in Web Applications.” Proceedings of the 2018 World Wide Web Conference on World Wide Web - WWW ’18. https://doi.org/10.1145/3178876.3185996.
Xu, Huan, Constantine Caramanis, and Shie Mannor. 2008. “Robustness and Regularization of Support Vector Machines.” arXiv. https://doi.org/10.48550/ARXIV.0803.3490.
Xu, Jian, Kuang-chih Lee, Wentong Li, Hang Qi, and Quan Lu. 2015. “Smart Pacing for Effective Online Ad Campaign Optimization.” Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2783258.2788615.
Xu, Jie, Benjamin S. Glicksberg, Chang Su, Peter Walker, Jiang Bian, and Fei Wang. 2019. “Federated Learning for Healthcare Informatics.” arXiv. https://doi.org/10.48550/ARXIV.1911.06270.
Xu, Jing, Minhui (Jason) Xue, and Stjepan Picek. 2021. “Explainability-Based Backdoor Attacks Against Graph Neural Networks.” Proceedings of the 3rd ACM Workshop on Wireless Security and Machine Learning, June. https://doi.org/10.1145/3468218.3469046.
Xu, Kaidi, Sijia Liu, Pin-Yu Chen, Mengshu Sun, Caiwen Ding, Bhavya Kailkhura, and Xue Lin. 2020. “Towards an Efficient and General Framework of Robust Training for Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2002.10947.
Xu, Kelvin, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhutdinov, Richard Zemel, and Yoshua Bengio. 2015. “Show, Attend and Tell: Neural Image Caption Generation with Visual Attention,” February. http://arxiv.org/abs/1502.03044v3.
Xu, Kun, Linfeng Song, Yansong Feng, Yan Song, and Dong Yu. 2020. “Coordinated Reasoning for Cross-Lingual Knowledge Graph Alignment.” arXiv. https://doi.org/10.48550/ARXIV.2001.08728.
Xu, Nuo, Pinghui Wang, Long Chen, Li Pan, Xiaoyan Wang, and Junzhou Zhao. 2020. “Distinguish Confusing Law Articles for Legal Judgment Prediction.” arXiv. https://doi.org/10.48550/ARXIV.2004.02557.
Xu, Qiangeng, Yin Zhou, Weiyue Wang, Charles R. Qi, and Dragomir Anguelov. 2021. “SPG: Unsupervised Domain Adaptation for 3D Object Detection via Semantic Point Generation.” arXiv. https://doi.org/10.48550/ARXIV.2108.06709.
Xu, Shangqing, and Chao Zhang. 2024. “Misconfidence-Based Demonstration Selection for LLM in-Context Learning.” arXiv. https://doi.org/10.48550/ARXIV.2401.06301.
Xu, Wei, Chris Callison-Burch, and Courtney Napoles. 2015. “Problems in Current Text Simplification Research: New Data Can Help.” Transactions of the Association for Computational Linguistics 3 (December). https://doi.org/10.1162/tacl_a_00139.
Xu, Weinan, Hengxu He, Minshi Tan, Yunming Li, Jun Lang, and Dongbai Guo. 2020. “Deep Interest with Hierarchical Attention Network for Click-Through Rate Prediction.” arXiv. https://doi.org/10.48550/ARXIV.2005.12981.
Xu, Xingqian, Jiayi Guo, Zhangyang Wang, Gao Huang, Irfan Essa, and Humphrey Shi. 2023. “Prompt-Free Diffusion: Taking "Text" Out of Text-to-Image Diffusion Models.” arXiv. https://doi.org/10.48550/ARXIV.2305.16223.
Xu, Yifan, Huapeng Wei, Minxuan Lin, Yingying Deng, Kekai Sheng, Mengdan Zhang, Fan Tang, Weiming Dong, Feiyue Huang, and Changsheng Xu. 2021. “Transformers in Computational Visual Media: A Survey.” Computational Visual Media 8 (October). https://doi.org/10.1007/s41095-021-0247-3.
Xu, Yuzhuang, Shuo Wang, Peng Li, Fuwen Luo, Xiaolong Wang, Weidong Liu, and Yang Liu. 2023. “Exploring Large Language Models for Communication Games: An Empirical Study on Werewolf.” arXiv. https://doi.org/10.48550/ARXIV.2309.04658.
Xu, Zhaozhuo, Zirui Liu, Beidi Chen, Yuxin Tang, Jue Wang, Kaixiong Zhou, Xia Hu, and Anshumali Shrivastava. 2023. “Compress, Then Prompt: Improving Accuracy-Efficiency Trade-Off of LLM Inference with Transferable Prompt.” arXiv. https://doi.org/10.48550/ARXIV.2305.11186.
Xu, Ziyun, Chengyu Wang, Minghui Qiu, Fuli Luo, Runxin Xu, Songfang Huang, and Jun Huang. 2022. “Making Pre-Trained Language Models End-to-End Few-Shot Learners with Contrastive Prompt Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2204.00166.
Xuanyuan, Han, Pietro Barbiero, Dobrik Georgiev, Lucie Charlotte Magister, and Pietro Lió. 2022. “Global Concept-Based Interpretability for Graph Neural Networks via Neuron Analysis.” arXiv. https://doi.org/10.48550/ARXIV.2208.10609.
Xue, Feng, Xiangnan He, Xiang Wang, Jiandong Xu, Kai Liu, and Richang Hong. 2018. “Deep Item-Based Collaborative Filtering for Top-n Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.1811.04392.
Xue, Hansheng, Luwei Yang, Wen Jiang, Yi Wei, Yi Hu, and Yu Lin. 2020. “Modeling Dynamic Heterogeneous Network for Link Prediction Using Hierarchical Attention with Temporal RNN.” arXiv. https://doi.org/10.48550/ARXIV.2004.01024.
Xue, Lanqing, Kaitao Song, Duocai Wu, Xu Tan, Nevin L. Zhang, Tao Qin, Wei-Qiang Zhang, and Tie-Yan Liu. 2021. “DeepRapper: Neural Rap Generation with Rhyme and Rhythm Modeling.” arXiv. https://doi.org/10.48550/ARXIV.2107.01875.
Yackulic, Charles B., Michael Dodrill, Maria Dzul, Jamie S. Sanderlin, and Janice A. Reid. 2020. “A Need for Speed in Bayesian Population Models: A Practical Guide to Marginalizing and Recovering Discrete Latent States.” Ecological Applications 30 (April). https://doi.org/10.1002/eap.2112.
Yahav, Eran. 2018. “From Programs to Interpretable Deep Models and Back.” Computer Aided Verification. https://doi.org/10.1007/978-3-319-96145-3_2.
Yahya, Ali, Adrian Li, Mrinal Kalakrishnan, Yevgen Chebotar, and Sergey Levine. 2017. “Collective Robot Reinforcement Learning with Distributed Asynchronous Guided Policy Search.” 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), September. https://doi.org/10.1109/iros.2017.8202141.
Yamada, Ikuya, Hiroyuki Shindo, and Yoshiyasu Takefuji. 2018. “Representation Learning of Entities and Documents from Knowledge Base Descriptions.” arXiv. https://doi.org/10.48550/ARXIV.1806.02960.
Yamamoto, Kohei, and Kurato Maeno. 2018. “PCAS: Pruning Channels with Attention Statistics for Deep Network Compression.” arXiv. https://doi.org/10.48550/ARXIV.1806.05382.
Yamamoto, Ryuichi, Martin Andrews, Michael Petrochuk, Wang Hy, Olga Vishnepolski, Matt Cooper, Kuan Chen, and Aleksas Pielikis. 2018. “R9y9/Wavenet_vocoder: V0.1.1 Release,” October. https://doi.org/10.5281/ZENODO.1472609.
Yan, Pei, Shunquan Tan, Miaohui Wang, and Jiwu Huang. 2023. “Prompt Engineering-Assisted Malware Dynamic Analysis Using GPT-4.” arXiv. https://doi.org/10.48550/ARXIV.2312.08317.
Yan, Rong, Jelena Tesic, and John R. Smith. 2007. “Model-Shared Subspace Boosting for Multi-Label Classification.” Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/1281192.1281281.
Yan, Shuyuan, Bolin Ding, Wei Guo, Jingren Zhou, Zhewei Wei, Xiaowei Jiang, and Sheng Xu. 2021. “FlashP.” Proceedings of the VLDB Endowment 14 (January). https://doi.org/10.14778/3446095.3446096.
Yan, Sijie, Yuanjun Xiong, and Dahua Lin. 2018. “Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1801.07455.
Yan, Siyuan, Zhen Yu, Xuelin Zhang, Dwarikanath Mahapatra, Shekhar S. Chandra, Monika Janda, Peter Soyer, and Zongyuan Ge. 2023. “Towards Trustable Skin Cancer Diagnosis via Rewriting Model’s Decision.” arXiv. https://doi.org/10.48550/ARXIV.2303.00885.
Yan, Xifeng, and Jiawei Han. 2003. “CloseGraph.” Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/956750.956784.
———. n.d. “gSpan: Graph-Based Substructure Pattern Mining.” 2002 IEEE International Conference on Data Mining, 2002. Proceedings. https://doi.org/10.1109/icdm.2002.1184038.
Yan, Yan, Yuxing Mao, and Bo Li. 2018. “SECOND: Sparsely Embedded Convolutional Detection.” Sensors 18 (October). https://doi.org/10.3390/s18103337.
Yan, Yukun, Daqi Zheng, Zhengdong Lu, and Sen Song. 2017. “Event Identification as a Decision Process with Non-Linear Representation of Text.” arXiv. https://doi.org/10.48550/ARXIV.1710.00969.
Yang, Bin, Lin Yang, Xiaochun Li, Wenhan Zhang, Hua Zhou, Yequn Zhang, Yongxiong Ren, and Yinbo Shi. 2019. “2-Bit Model Compression of Deep Convolutional Neural Network on ASIC Engine for Image Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.1905.03362.
Yang, Bishan, and Tom Mitchell. 2019. “Leveraging Knowledge Bases in LSTMs for Improving Machine Reading.” arXiv. https://doi.org/10.48550/ARXIV.1902.09091.
Yang, Bishan, Wen-tau Yih, Xiaodong He, Jianfeng Gao, and Li Deng. 2014. “Embedding Entities and Relations for Learning and Inference in Knowledge Bases.” arXiv. https://doi.org/10.48550/ARXIV.1412.6575.
Yang, Bo, Xiao Fu, Nicholas D. Sidiropoulos, and Mingyi Hong. 2016. “Towards k-Means-Friendly Spaces: Simultaneous Deep Learning and Clustering.” arXiv. https://doi.org/10.48550/ARXIV.1610.04794.
Yang, Ceyuan, Zhirong Wu, Bolei Zhou, and Stephen Lin. 2021. “Instance Localization for Self-Supervised Detection Pretraining.” arXiv. https://doi.org/10.48550/ARXIV.2102.08318.
Yang, Chenglin, Lingxi Xie, Chi Su, and Alan L. Yuille. 2018. “Snapshot Distillation: Teacher-Student Optimization in One Generation.” arXiv. https://doi.org/10.48550/ARXIV.1812.00123.
Yang, Chengrun, Yuji Akimoto, Dae Won Kim, and Madeleine Udell. 2019. “OBOE.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330909.
Yang, Guandao, Xun Huang, Zekun Hao, Ming-Yu Liu, Serge Belongie, and Bharath Hariharan. 2019. “PointFlow: 3D Point Cloud Generation with Continuous Normalizing Flows.” arXiv. https://doi.org/10.48550/ARXIV.1906.12320.
Yang, Guanglei, Hao Tang, Mingli Ding, Nicu Sebe, and Elisa Ricci. 2021. “Transformer-Based Attention Networks for Continuous Pixel-Wise Prediction.” arXiv. https://doi.org/10.48550/ARXIV.2103.12091.
Yang, Haichuan, Yuhao Zhu, and Ji Liu. 2018a. “ECC: Platform-Independent Energy-Constrained Deep Neural Network Compression via a Bilinear Regression Model.” arXiv. https://doi.org/10.48550/ARXIV.1812.01803.
———. 2018b. “Energy-Constrained Compression for Deep Neural Networks via Weighted Sparse Projection and Layer Input Masking.” arXiv. https://doi.org/10.48550/ARXIV.1806.04321.
Yang, Haode, Parth Gupta, Roberto Fernández Galán, Dan Bu, and Dongmei Jia. 2021. “Seasonal Relevance in e-Commerce Search.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3481951.
Yang, Jihan, Shaoshuai Shi, Zhe Wang, Hongsheng Li, and Xiaojuan Qi. 2021. “ST3D: Self-Training for Unsupervised Domain Adaptation on 3D Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2103.05346.
Yang, Jingzhou, and Lei He. 2020. “Towards Universal Text-to-Speech.” Interspeech 2020, October. https://doi.org/10.21437/interspeech.2020-1590.
Yang, Ji, Xinyang Yi, Derek Zhiyuan Cheng, Lichan Hong, Yang Li, Simon Xiaoming Wang, Taibai Xu, and Ed H. Chi. 2020. “Mixed Negative Sampling for Learning Two-Tower Neural Networks in Recommendations.” Companion Proceedings of the Web Conference 2020, April. https://doi.org/10.1145/3366424.3386195.
Yang, Kaiyu, and Fanhuai Shi. 2023. “Medium- and Long-Term Load Forecasting for Power Plants Based on Causal Inference and Informer.” Applied Sciences 13 (June). https://doi.org/10.3390/app13137696.
Yang, Ling, Zhilong Zhang, Yang Song, Shenda Hong, Runsheng Xu, Yue Zhao, Wentao Zhang, Bin Cui, and Ming-Hsuan Yang. 2022. “Diffusion Models: A Comprehensive Survey of Methods and Applications.” arXiv. https://doi.org/10.48550/ARXIV.2209.00796.
Yang, Li, and Abdallah Shami. 2022. “IoT Data Analytics in Dynamic Environments: From an Automated Machine Learning Perspective.” Engineering Applications of Artificial Intelligence 116 (November). https://doi.org/10.1016/j.engappai.2022.105366.
Yang, Mouxing, Zhenyu Huang, and Xi Peng. 2024. “Robust Object Re-Identification with Coupled Noisy Labels.” International Journal of Computer Vision, February. https://doi.org/10.1007/s11263-024-01997-w.
Yang, Nan, and Renyu Zhang. 2022. “Comparative Statics Analysis of an Inventory Management Model with Dynamic Pricing, Market Environment Fluctuation, and Delayed Differentiation.” Production and Operations Management 31 (January). https://doi.org/10.1111/poms.13538.
Yang, Qiang, Yang Liu, Tianjian Chen, and Yongxin Tong. 2019. “Federated Machine Learning: Concept and Applications.” arXiv. https://doi.org/10.48550/ARXIV.1902.04885.
Yang, Ruichao, Xiting Wang, Yiqiao Jin, Chaozhuo Li, Jianxun Lian, and Xing Xie. 2022. “Reinforcement Subgraph Reasoning for Fake News Detection.” Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3534678.3539277.
Yang, Rui, Li Fang, and Yi Zhou. 2023. “Can Text-Based Knowledge Graph Completion Benefit from Zero-Shot Large Language Models?” arXiv. https://doi.org/10.48550/ARXIV.2310.08279.
Yang, Sherry, Ofir Nachum, Yilun Du, Jason Wei, Pieter Abbeel, and Dale Schuurmans. 2023. “Foundation Models for Decision Making: Problems, Methods, and Opportunities.” arXiv. https://doi.org/10.48550/ARXIV.2303.04129.
Yang, Tien-Ju, Yu-Hsin Chen, and Vivienne Sze. 2016. “Designing Energy-Efficient Convolutional Neural Networks Using Energy-Aware Pruning.” arXiv. https://doi.org/10.48550/ARXIV.1611.05128.
Yang, Xianjun, Yan Li, Xinlu Zhang, Haifeng Chen, and Wei Cheng. 2023. “Exploring the Limits of ChatGPT for Query or Aspect-Based Text Summarization.” arXiv. https://doi.org/10.48550/ARXIV.2302.08081.
Yang, Xiaocui, Shi Feng, Daling Wang, Sun Qi, Wenfang Wu, Yifei Zhang, Pengfei Hong, and Soujanya Poria. 2023. “Few-Shot Joint Multimodal Aspect-Sentiment Analysis Based on Generative Multimodal Prompt.” arXiv. https://doi.org/10.48550/ARXIV.2305.10169.
Yang, Yang, Yuhong Xu, Chunping Wang, Yizhou Sun, Fei Wu, Yueting Zhuang, and Ming Gu. 2019. “Understanding Default Behavior in Online Lending.” Proceedings of the 28th ACM International Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3357384.3358052.
Yang, Yifan, Qijing Huang, Bichen Wu, Tianjun Zhang, Liang Ma, Giulio Gambardella, Michaela Blott, et al. 2019. “Synetgy.” Proceedings of the 2019 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, February. https://doi.org/10.1145/3289602.3293902.
Yang, Yingguang, Renyu Yang, Yangyang Li, Kai Cui, Zhiqin Yang, Yue Wang, Jie Xu, and Haiyong Xie. 2023. “<Scp>RoSGAS</Scp> : Adaptive Social Bot Detection with Reinforced Self-Supervised GNN Architecture Search.” ACM Transactions on the Web 17 (May). https://doi.org/10.1145/3572403.
Yang, Ze, Linjun Shou, Ming Gong, Wutao Lin, and Daxin Jiang. 2019. “Model Compression with Multi-Task Knowledge Distillation for Web-Scale Question Answering System.” arXiv. https://doi.org/10.48550/ARXIV.1904.09636.
Yao, Bingsheng, Guiming Chen, Ruishi Zou, Yuxuan Lu, Jiachen Li, Shao Zhang, Sijia Liu, James Hendler, and Dakuo Wang. 2023. “More Samples or More Prompt Inputs? Exploring Effective in-Context Sampling for LLM Few-Shot Prompt Engineering.” arXiv. https://doi.org/10.48550/ARXIV.2311.09782.
Yao, Hantao, Rui Zhang, and Changsheng Xu. 2023. “Visual-Language Prompt Tuning with Knowledge-Guided Context Optimization.” arXiv. https://doi.org/10.48550/ARXIV.2303.13283.
Yao, Hongwei, Jian Lou, and Zhan Qin. 2023. “PoisonPrompt: Backdoor Attack on Prompt-Based Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2310.12439.
Yao, Hongwei, Jian Lou, Kui Ren, and Zhan Qin. 2023. “PromptCARE: Prompt Copyright Protection by Watermark Injection and Verification.” arXiv. https://doi.org/10.48550/ARXIV.2308.02816.
Yao, Huaxiu, Ying Wei, Junzhou Huang, and Zhenhui Li. 2019. “Hierarchically Structured Meta-Learning.” arXiv. https://doi.org/10.48550/ARXIV.1905.05301.
Yao, Jie-En, Li-Yuan Tsao, Yi-Chen Lo, Roy Tseng, Chia-Che Chang, and Chun-Yi Lee. 2023. “Local Implicit Normalizing Flow for Arbitrary-Scale Image Super-Resolution.” arXiv. https://doi.org/10.48550/ARXIV.2303.05156.
Yao, Liang, Chengsheng Mao, and Yuan Luo. 2019. “KG-BERT: BERT for Knowledge Graph Completion.” arXiv. https://doi.org/10.48550/ARXIV.1909.03193.
Yao, Quanming, Mengshuo Wang, Yuqiang Chen, Wenyuan Dai, Yu-Feng Li, Wei-Wei Tu, Qiang Yang, and Yang Yu. 2018. “Taking Human Out of Learning Applications: A Survey on Automated Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1810.13306.
Yao, Tiansheng, Xinyang Yi, Derek Zhiyuan Cheng, Felix Yu, Ting Chen, Aditya Menon, Lichan Hong, et al. 2020. “Self-Supervised Learning for Large-Scale Item Recommendations.” arXiv. https://doi.org/10.48550/ARXIV.2007.12865.
Yao, Yuan, Deming Ye, Peng Li, Xu Han, Yankai Lin, Zhenghao Liu, Zhiyuan Liu, Lixin Huang, Jie Zhou, and Maosong Sun. 2019. “DocRED: A Large-Scale Document-Level Relation Extraction Dataset.” arXiv. https://doi.org/10.48550/ARXIV.1906.06127.
Yao, Yuling, Gregor Pirš, Aki Vehtari, and Andrew Gelman. 2022. “Bayesian Hierarchical Stacking: Some Models Are (Somewhere) Useful.” Bayesian Analysis 17 (December). https://doi.org/10.1214/21-ba1287.
Yao, Yuling, Aki Vehtari, Daniel Simpson, and Andrew Gelman. 2018a. “Yes, but Did It Work?: Evaluating Variational Inference.” arXiv. https://doi.org/10.48550/ARXIV.1802.02538.
———. 2018b. “Using Stacking to Average Bayesian Predictive Distributions (with Discussion).” Bayesian Analysis 13 (September). https://doi.org/10.1214/17-ba1091.
Yarowsky, David. 1995. “Unsupervised Word Sense Disambiguation Rivaling Supervised Methods.” Proceedings of the 33rd Annual Meeting on Association for Computational Linguistics -. https://doi.org/10.3115/981658.981684.
Ye, Jianbo, Xin Lu, Zhe Lin, and James Z. Wang. 2018. “Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution Layers.” arXiv. https://doi.org/10.48550/ARXIV.1802.00124.
Ye, Qinghao, Haiyang Xu, Guohai Xu, Jiabo Ye, Ming Yan, Yiyang Zhou, Junyang Wang, et al. 2023. “mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality.” arXiv. https://doi.org/10.48550/ARXIV.2304.14178.
Ye, Qinyuan, Bill Yuchen Lin, and Xiang Ren. 2021. “CrossFit: A Few-Shot Learning Challenge for Cross-Task Generalization in NLP.” arXiv. https://doi.org/10.48550/ARXIV.2104.08835.
Ye, Rui, Xin Li, Yujie Fang, Hongyu Zang, and Mingzhong Wang. 2019. “A Vectorized Relational Graph Convolutional Network for Multi-Relational Network Alignment.” Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, August. https://doi.org/10.24963/ijcai.2019/574.
Ye, Shaokai, Tianyun Zhang, Kaiqi Zhang, Jiayu Li, Jiaming Xie, Yun Liang, Sijia Liu, Xue Lin, and Yanzhi Wang. 2018. “A Unified Framework of DNN Weight Pruning and Weight Clustering/Quantization Using ADMM.” arXiv. https://doi.org/10.48550/ARXIV.1811.01907.
Ye, Zhi-Sheng, Yili Hong, and Yimeng Xie. 2013. “How Do Heterogeneities in Operating Environments Affect Field Failure Predictions and Test Planning?” The Annals of Applied Statistics 7 (December). https://doi.org/10.1214/13-aoas666.
Yeaton, Anna, Rahul G. Krishnan, Rebecca Mieloszyk, David Alvarez-Melis, and Grace Huynh. 2022. “Hierarchical Optimal Transport for Comparing Histopathology Datasets.” arXiv. https://doi.org/10.48550/ARXIV.2204.08324.
Yeh, Chih-Kuan, Cheng-Yu Hsieh, Arun Sai Suggala, David I. Inouye, and Pradeep Ravikumar. 2019. “On the (in)fidelity and Sensitivity for Explanations.” arXiv. https://doi.org/10.48550/ARXIV.1901.09392.
Yeh, Hui-Syuan, Thomas Lavergne, and Pierre Zweigenbaum. 2022. “Decorate the Examples: A Simple Method of Prompt Design for Biomedical Relation Extraction.” arXiv. https://doi.org/10.48550/ARXIV.2204.10360.
Yeo, Yee Hui, Jamil S. Samaan, Wee Han Ng, Peng-Sheng Ting, Hirsh Trivedi, Aarshi Vipani, Walid Ayoub, et al. 2023. “Assessing the Performance of ChatGPT in Answering Questions Regarding Cirrhosis and Hepatocellular Carcinoma,” February. https://doi.org/10.1101/2023.02.06.23285449.
Yeung, Serena, Alireza Fathi, and Li Fei-Fei. 2014. “VideoSET: Video Summary Evaluation Through Text.” arXiv. https://doi.org/10.48550/ARXIV.1406.5824.
Yi, Dong, Zhen Lei, Shengcai Liao, and Stan Z. Li. 2014. “Learning Face Representation from Scratch.” arXiv. https://doi.org/10.48550/ARXIV.1411.7923.
Yi, Xing, Liangjie Hong, Erheng Zhong, Nanthan Nan Liu, and Suju Rajan. 2014. “Beyond Clicks.” Proceedings of the 8th ACM Conference on Recommender Systems, October. https://doi.org/10.1145/2645710.2645724.
Yikun, Ban, Liu Xin, Huang Ling, Duan Yitao, Liu Xue, and Xu Wei. 2019. “No Place to Hide: Catching Fraudulent Entities in Tensors.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313403.
Yim, Junho, Donggyu Joo, Jihoon Bae, and Junmo Kim. 2017. “A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July. https://doi.org/10.1109/cvpr.2017.754.
Yin, Aoxiong, Zhou Zhao, Weike Jin, Meng Zhang, Xingshan Zeng, and Xiaofei He. 2022. “MLSLT: Towards Multilingual Sign Language Translation.” 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June. https://doi.org/10.1109/cvpr52688.2022.00505.
Yin, Dawei, Yuening Hu, Jiliang Tang, Tim Daly, Mianwei Zhou, Hua Ouyang, Jianhui Chen, et al. 2016. “Ranking Relevance in Yahoo Search.” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2939672.2939677.
Yin, Fan, Weining Shen, and Carter T. Butts. 2022. “Finite Mixtures of ERGMs for Modeling Ensembles of Networks.” Bayesian Analysis 17 (December). https://doi.org/10.1214/21-ba1298.
Yin, Junbo, Dingfu Zhou, Liangjun Zhang, Jin Fang, Cheng-Zhong Xu, Jianbing Shen, and Wenguan Wang. 2022. “ProposalContrast: Unsupervised Pre-Training for LiDAR-Based 3D Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2207.12654.
Yin, Kangxue, Zhiqin Chen, Hui Huang, Daniel Cohen-Or, and Hao Zhang. 2019. “LOGAN.” ACM Transactions on Graphics 38 (November). https://doi.org/10.1145/3355089.3356494.
Yin, Pengcheng, Graham Neubig, Wen-tau Yih, and Sebastian Riedel. 2020. “TaBERT: Pretraining for Joint Understanding of Textual and Tabular Data.” arXiv. https://doi.org/10.48550/ARXIV.2005.08314.
Yin, Rujie. 2016. “Content Aware Neural Style Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1601.04568.
Yin, Tianwei, Xingyi Zhou, and Philipp Krähenbühl. 2021. “Multimodal Virtual Point 3D Detection.” arXiv. https://doi.org/10.48550/ARXIV.2111.06881.
Ying, Rex, Ruining He, Kaifeng Chen, Pong Eksombatchai, William L. Hamilton, and Jure Leskovec. 2018. “Graph Convolutional Neural Networks for Web-Scale Recommender Systems.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3219890.
Yoneyama, Reo, Yi-Chiao Wu, and Tomoki Toda. 2022. “Source-Filter HiFi-GAN: Fast and Pitch Controllable High-Fidelity Neural Vocoder.” arXiv. https://doi.org/10.48550/ARXIV.2210.15533.
Yoo, Sanghyun, Young-Seok Kim, Kang Hyun Lee, Kuhwan Jeong, Junhwi Choi, Hoshik Lee, and Young Sang Choi. 2020. “Graph-Aware Transformer: Is Attention All Graphs Need?” arXiv. https://doi.org/10.48550/ARXIV.2006.05213.
Yoon, KiJung, Renjie Liao, Yuwen Xiong, Lisa Zhang, Ethan Fetaya, Raquel Urtasun, Richard Zemel, and Xaq Pitkow. 2018. “Inference in Probabilistic Graphical Models by Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1803.07710.
Yoon, Sung Whan, Jun Seo, and Jaekyun Moon. 2019. “TapNet: Neural Network Augmented with Task-Adaptive Projection for Few-Shot Learning.” arXiv. https://doi.org/10.48550/ARXIV.1905.06549.
Yoon, Susik, Jae-Gil Lee, and Byung Suk Lee. 2019. “NETS.” Proceedings of the VLDB Endowment 12 (July). https://doi.org/10.14778/3342263.3342269.
Yosinski, Jason, Jeff Clune, Anh Nguyen, Thomas Fuchs, and Hod Lipson. 2015. “Understanding Neural Networks Through Deep Visualization.” arXiv. https://doi.org/10.48550/ARXIV.1506.06579.
You, Jiaxuan, Tianyu Du, and Jure Leskovec. 2022. “ROLAND: Graph Learning Framework for Dynamic Graphs.” Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3534678.3539300.
You, Yurong, Katie Z Luo, Xiangyu Chen, Junan Chen, Wei-Lun Chao, Wen Sun, Bharath Hariharan, Mark Campbell, and Kilian Q. Weinberger. 2022. “Hindsight Is 20/20: Leveraging Past Traversals to Aid 3D Perception.” arXiv. https://doi.org/10.48550/ARXIV.2203.11405.
Younes, Laurent. 1989. “Parametric Inference for Imperfectly Observed Gibbsian Fields.” Probability Theory and Related Fields 82 (August). https://doi.org/10.1007/bf00341287.
Yu, Donghan, Yiming Yang, Ruohong Zhang, and Yuexin Wu. 2021. “Knowledge Embedding Based Graph Convolutional Network.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3449925.
Yu, Fisher, Ari Seff, Yinda Zhang, Shuran Song, Thomas Funkhouser, and Jianxiong Xiao. 2015. “LSUN: Construction of a Large-Scale Image Dataset Using Deep Learning with Humans in the Loop.” arXiv. https://doi.org/10.48550/ARXIV.1506.03365.
Yu, Guang, Siqi Wang, Zhiping Cai, Xinwang Liu, Chuanfu Xu, and Chengkun Wu. 2021. “Deep Anomaly Discovery from Unlabeled Videos via Normality Advantage and Self-Paced Refinement.” arXiv. https://doi.org/10.48550/ARXIV.2108.01975.
Yu, Jiahui, Chung-Cheng Chiu, Bo Li, Shuo-yiin Chang, Tara N. Sainath, Yanzhang He, Arun Narayanan, et al. 2020. “FastEmit: Low-Latency Streaming ASR with Sequence-Level Emission Regularization.” arXiv. https://doi.org/10.48550/ARXIV.2010.11148.
Yu, Jiahui, Wei Han, Anmol Gulati, Chung-Cheng Chiu, Bo Li, Tara N. Sainath, Yonghui Wu, and Ruoming Pang. 2020. “Dual-Mode ASR: Unify and Improve Streaming ASR with Full-Context Modeling.” arXiv. https://doi.org/10.48550/ARXIV.2010.06030.
Yu, Jianke, Hanchen Wang, Xiaoyang Wang, Zhao Li, Lu Qin, Wenjie Zhang, Jian Liao, and Ying Zhang. 2023. “Group-Based Fraud Detection Network on e-Commerce Platforms.” Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/3580305.3599836.
Yu, Rose, Huida Qiu, Zhen Wen, Ching-Yung Lin, and Yan Liu. 2016. “A Survey on Social Media Anomaly Detection.” arXiv. https://doi.org/10.48550/ARXIV.1601.01102.
Yu, Wenchao, Wei Cheng, Charu C. Aggarwal, Kai Zhang, Haifeng Chen, and Wei Wang. 2018. “NetWalk.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220024.
Yu, Yue, Jie Chen, Tian Gao, and Mo Yu. 2019. “DAG-GNN: DAG Structure Learning with Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1904.10098.
Yu, Zhiwen, Xingshe Zhou, Yanbin Hao, and Jianhua Gu. 2006. “TV Program Recommendation for Multiple Viewers Based on User Profile Merging.” User Modeling and User-Adapted Interaction 16 (March). https://doi.org/10.1007/s11257-006-9005-6.
Yuan, Fajie, Xiangnan He, Haochuan Jiang, Guibing Guo, Jian Xiong, Zhezhao Xu, and Yilin Xiong. 2019. “Future Data Helps Training: Modeling Future Contexts for Session-Based Recommendation.” arXiv. https://doi.org/10.48550/ARXIV.1906.04473.
———. 2020. “Future Data Helps Training: Modeling Future Contexts for Session-Based Recommendation.” Proceedings of The Web Conference 2020, April. https://doi.org/10.1145/3366423.3380116.
Yuan, Fajie, Guoxiao Zhang, Alexandros Karatzoglou, Joemon Jose, Beibei Kong, and Yudong Li. 2021. “One Person, One Model, One World.” Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3404835.3462884.
Yuan, Haoqi, Chi Zhang, Hongcheng Wang, Feiyang Xie, Penglin Cai, Hao Dong, and Zongqing Lu. 2023. “Skill Reinforcement Learning and Planning for Open-World Long-Horizon Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2303.16563.
Yuan, Hao, Jiliang Tang, Xia Hu, and Shuiwang Ji. 2020. “XGNN: Towards Model-Level Explanations of Graph Neural Networks.” Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, August. https://doi.org/10.1145/3394486.3403085.
Yuan, Lifan, Yangyi Chen, Ganqu Cui, Hongcheng Gao, Fangyuan Zou, Xingyi Cheng, Heng Ji, Zhiyuan Liu, and Maosong Sun. 2023. “Revisiting Out-of-Distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations.” arXiv. https://doi.org/10.48550/ARXIV.2306.04618.
Yuan, Li, Francis E. H. Tay, Guilin Li, Tao Wang, and Jiashi Feng. 2019. “Revisiting Knowledge Distillation via Label Smoothing Regularization.” arXiv. https://doi.org/10.48550/ARXIV.1909.11723.
Yuan, Xiao-Tong, and Tong Zhang. 2011. “Truncated Power Method for Sparse Eigenvalue Problems.” arXiv. https://doi.org/10.48550/ARXIV.1112.2679.
Yuan, Yijun, and Andreas Nuechter. 2023. “Uni-Fusion: Universal Continuous Mapping.” arXiv. https://doi.org/10.48550/ARXIV.2303.12678.
Yuan, Zheng, Fajie Yuan, Yu Song, Youhua Li, Junchen Fu, Fei Yang, Yunzhu Pan, and Yongxin Ni. 2023. “Where to Go Next for Recommender Systems? ID- Vs. Modality-Based Recommender Models Revisited.” Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3539618.3591932.
Yuan, Zheng, Hongyi Yuan, Chuanqi Tan, Wei Wang, and Songfang Huang. 2023. “How Well Do Large Language Models Perform in Arithmetic Tasks?” arXiv. https://doi.org/10.48550/ARXIV.2304.02015.
Yuan, Zhenlong, Yongqiang Lu, Zhaoguo Wang, and Yibo Xue. 2014. “Droid-Sec.” Proceedings of the 2014 ACM Conference on SIGCOMM, August. https://doi.org/10.1145/2619239.2631434.
Yue, Li, Zhao Weibin, and Shang Lin. 2019. “Really Should We Pruning After Model Be Totally Trained? Pruning Based on a Small Amount of Training.” arXiv. https://doi.org/10.48550/ARXIV.1901.08455.
Zador, Anthony M. 2019. “A Critique of Pure Learning and What Artificial Neural Networks Can Learn from Animal Brains.” Nature Communications 10 (August). https://doi.org/10.1038/s41467-019-11786-6.
Zafrir, Ofir, Guy Boudoukh, Peter Izsak, and Moshe Wasserblat. 2019. “Q8BERT: Quantized 8Bit BERT.” 2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing - NeurIPS Edition (EMC2-NIPS), December. https://doi.org/10.1109/emc2-nips53020.2019.00016.
Zagoruyko, Sergey, and Nikos Komodakis. 2016a. “Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer.” arXiv. https://doi.org/10.48550/ARXIV.1612.03928.
———. 2016b. “Wide Residual Networks.” arXiv. https://doi.org/10.48550/ARXIV.1605.07146.
Zaheer, Manzil, Satwik Kottur, Siamak Ravanbakhsh, Barnabas Poczos, Ruslan Salakhutdinov, and Alexander Smola. 2017. “Deep Sets.” arXiv. https://doi.org/10.48550/ARXIV.1703.06114.
Zähle, M. 1998. “Integration with Respect to Fractal Functions and Stochastic Calculus. i.” Probability Theory and Related Fields 111 (July). https://doi.org/10.1007/s004400050171.
Zambaldi, Vinicius, David Raposo, Adam Santoro, Victor Bapst, Yujia Li, Igor Babuschkin, Karl Tuyls, et al. 2018. “Relational Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1806.01830.
Zamini, Mohamad, and Eunjin Kim. 2022. “A Survey on Computational Intelligence-Based Transfer Learning.” arXiv. https://doi.org/10.48550/ARXIV.2206.10593.
Zanardini, Damiano, and Emilio Serrano. 2024. “Introducing New Node Prediction in Graph Mining: Predicting All Links from Isolated Nodes with Graph Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2401.05468.
Zandvoort, Daphne van, Laura Wiersema, Tom Huibers, Sandra van Dulmen, and Sjaak Brinkkemper. 2023. “Enhancing Summarization Performance Through Transformer-Based Prompt Engineering in Automated Medical Reporting.” arXiv. https://doi.org/10.48550/ARXIV.2311.13274.
Zantedeschi, Valentina, Matt J. Kusner, and Vlad Niculae. 2020. “Learning Binary Decision Trees by Argmin Differentiation.” arXiv. https://doi.org/10.48550/ARXIV.2010.04627.
Zappone, Alessio, Marco Di Renzo, and Mérouane Debbah. 2019. “Wireless Networks Design in the Era of Deep Learning: Model-Based, AI-Based, or Both?” arXiv. https://doi.org/10.48550/ARXIV.1902.02647.
Zaremba, Wojciech, Ilya Sutskever, and Oriol Vinyals. 2014. “Recurrent Neural Network Regularization.” arXiv. https://doi.org/10.48550/ARXIV.1409.2329.
Zee, Field G. Van, and Robert A. van de Geijn. 2015. “BLIS: A Framework for Rapidly Instantiating BLAS Functionality.” ACM Transactions on Mathematical Software 41 (June). https://doi.org/10.1145/2764454.
Zeiler, Matthew D. 2012. “ADADELTA: An Adaptive Learning Rate Method.” arXiv. https://doi.org/10.48550/ARXIV.1212.5701.
Zellers, Rowan, Ari Holtzman, Yonatan Bisk, Ali Farhadi, and Yejin Choi. 2019. “HellaSwag: Can a Machine Really Finish Your Sentence?” arXiv. https://doi.org/10.48550/ARXIV.1905.07830.
Zellers, Rowan, Ari Holtzman, Hannah Rashkin, Yonatan Bisk, Ali Farhadi, Franziska Roesner, and Yejin Choi. 2019. “Defending Against Neural Fake News.” arXiv. https://doi.org/10.48550/ARXIV.1905.12616.
Zemplenyi, Michele, and Jeffrey W. Miller. 2023. “Bayesian Optimal Experimental Design for Inferring Causal Structure.” Bayesian Analysis 18 (September). https://doi.org/10.1214/22-ba1335.
Zeng, Kaisheng, Zhenhao Dong, Lei Hou, Yixin Cao, Minghao Hu, Jifan Yu, Xin Lv, et al. 2022. “Interactive Contrastive Learning for Self-Supervised Entity Alignment.” Proceedings of the 31st ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3511808.3557364.
Zeng, Weixin, Xiang Zhao, Wei Wang, Jiuyang Tang, and Zhen Tan. 2020. “Degree-Aware Alignment for Entities in Tail.” arXiv. https://doi.org/10.48550/ARXIV.2005.12132.
Zeng, Wenyuan, Wenjie Luo, Sanja Fidler, and Raquel Urtasun. 2016. “Efficient Summarization with Read-Again and Copy Mechanism.” arXiv. https://doi.org/10.48550/ARXIV.1611.03382.
Zeng, Yankai, Abhiramon Rajasekharan, Parth Padalkar, Kinjal Basu, Joaquín Arias, and Gopal Gupta. 2023. “Automated Interactive Domain-Specific Conversational Agents That Understand Human Dialogs.” arXiv. https://doi.org/10.48550/ARXIV.2303.08941.
Zeng, Zexian, Yu Deng, Xiaoyu Li, Tristan Naumann, and Yuan Luo. 2018. “Natural Language Processing for EHR-Based Computational Phenotyping.” arXiv. https://doi.org/10.48550/ARXIV.1806.04820.
Zeng, Zhen, Jianzong Wang, Ning Cheng, Tian Xia, and Jing Xiao. 2020. “AlignTTS: Efficient Feed-Forward Text-to-Speech System Without Explicit Alignment.” arXiv. https://doi.org/10.48550/ARXIV.2003.01950.
Zha, Hanwen, Zhiyu Chen, and Xifeng Yan. 2021. “Inductive Relation Prediction by BERT.” arXiv. https://doi.org/10.48550/ARXIV.2103.07102.
Zhai, Shuangfei, Walter Talbott, Nitish Srivastava, Chen Huang, Hanlin Goh, Ruixiang Zhang, and Josh Susskind. 2021. “An Attention Free Transformer.” arXiv. https://doi.org/10.48550/ARXIV.2105.14103.
Zhang, Baichuan, and Mohammad Al Hasan. 2017. “Name Disambiguation in Anonymized Graphs Using Network Embedding.” Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3132847.3132873.
Zhang, Biao, Barry Haddow, and Alexandra Birch. 2023. “Prompting Large Language Model for Machine Translation: A Case Study.” arXiv. https://doi.org/10.48550/ARXIV.2301.07069.
Zhang, Bosen, Haiyan Huang, Laura E. Tibbs-Cortes, Adam Vanous, Zhiwu Zhang, Karen Sanguinet, Kimberly A. Garland-Campbell, Jianming Yu, and Xianran Li. 2023. “Streamline Unsupervised Machine Learning to Survey and Graph Indel-Based Haplotypes from Pan-Genomes,” February. https://doi.org/10.1101/2023.02.11.527743.
Zhang, Chen, Qiuchi Li, and Dawei Song. 2019. “Aspect-Based Sentiment Classification with Aspect-Specific Graph Convolutional Networks.” arXiv. https://doi.org/10.48550/ARXIV.1909.03477.
Zhang, Chenrui, Lin Liu, Jinpeng Wang, Chuyuan Wang, Xiao Sun, Hongyu Wang, and Mingchen Cai. 2023. “PREFER: Prompt Ensemble Learning via Feedback-Reflect-Refine.” arXiv. https://doi.org/10.48550/ARXIV.2308.12033.
Zhang, Chenshuang, Chaoning Zhang, Taegoo Kang, Donghun Kim, Sung-Ho Bae, and In So Kweon. 2023. “Attack-SAM: Towards Attacking Segment Anything Model with Adversarial Examples.” arXiv. https://doi.org/10.48550/ARXIV.2305.00866.
Zhang, Chenshuang, Chaoning Zhang, Mengchun Zhang, and In So Kweon. 2023. “Text-to-Image Diffusion Models in Generative AI: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2303.07909.
Zhang, Chen, Di Wu, Jiayu Sun, Guangyu Sun, Guojie Luo, and Jason Cong. 2016. “Energy-Efficient CNN Implementation on a Deeply Pipelined FPGA Cluster.” Proceedings of the 2016 International Symposium on Low Power Electronics and Design, August. https://doi.org/10.1145/2934583.2934644.
Zhang, Chenyuan, Hao Liu, Jiutian Zeng, Kejing Yang, Yuhong Li, and Hui Li. 2023. “Prompt-Enhanced Software Vulnerability Detection Using ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2308.12697.
Zhang, Chong, He Zhu, Xingyu Peng, Junran Wu, and Ke Xu. 2021. “Hierarchical Information Matters: Text Classification via Tree Based Graph Neural Network.” arXiv. https://doi.org/10.48550/ARXIV.2110.02047.
Zhang, Chuang, Qizhou Wang, Tengfei Liu, Xun Lu, Jin Hong, Bo Han, and Chen Gong. 2021. “Fraud Detection Under Multi-Sourced Extremely Noisy Annotations.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3482433.
Zhang, Daokun, Jie Yin, Xingquan Zhu, and Chengqi Zhang. 2018. “Network Representation Learning: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.1801.05852.
Zhang, Dinghuai, Tianyuan Zhang, Yiping Lu, Zhanxing Zhu, and Bin Dong. 2019. “You Only Propagate Once: Accelerating Adversarial Training via Maximal Principle.” arXiv. https://doi.org/10.48550/ARXIV.1905.00877.
Zhang, Haifeng, Weinan Zhang, Yifei Rong, Kan Ren, Wenxin Li, and Jun Wang. 2017. “Managing Risk of Bidding in Display Advertising.” Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/3018661.3018701.
Zhang, Han, Ian Goodfellow, Dimitris Metaxas, and Augustus Odena. 2018. “Self-Attention Generative Adversarial Networks.” arXiv. https://doi.org/10.48550/ARXIV.1805.08318.
Zhang, Hantian, Jerry Li, Kaan Kara, Dan Alistarh, Ji Liu, and Ce Zhang. 2016. “The ZipML Framework for Training Models with End-to-End Low Precision: The Cans, the Cannots, and a Little Bit of Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.05402.
Zhang, Han, Songlin Wang, Kang Zhang, Zhiling Tang, Yunjiang Jiang, Yun Xiao, Weipeng Yan, and Wen-Yun Yang. 2020. “Towards Personalized and Semantic Retrieval: An End-to-End Solution for e-Commerce Search via Embedding Learning.” arXiv. https://doi.org/10.48550/ARXIV.2006.02282.
Zhang, Hao, Jae Ro, and Richard Sproat. 2020. “Semi-Supervised URL Segmentation with Recurrent Neural Networks Pre-Trained on Knowledge Graph Entities.” Proceedings of the 28th International Conference on Computational Linguistics. https://doi.org/10.18653/v1/2020.coling-main.411.
Zhang, Haoyu, Jianjun Xu, and Ji Wang. 2019. “Pretraining-Based Natural Language Generation for Text Summarization.” arXiv. https://doi.org/10.48550/ARXIV.1902.09243.
Zhang, He, Bang Wu, Xingliang Yuan, Shirui Pan, Hanghang Tong, and Jian Pei. 2022. “Trustworthy Graph Neural Networks: Aspects, Methods and Trends.” arXiv. https://doi.org/10.48550/ARXIV.2205.07424.
Zhang, Hongming, Liwei Qiu, Lingling Yi, and Yangqiu Song. 2018. “Scalable Multiplex Network Embedding.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/428.
Zhang, Hongyang, Yaodong Yu, Jiantao Jiao, Eric P. Xing, Laurent El Ghaoui, and Michael I. Jordan. 2019. “Theoretically Principled Trade-Off Between Robustness and Accuracy.” arXiv. https://doi.org/10.48550/ARXIV.1901.08573.
Zhang, Jian, Zoubin Ghahramani, and Yiming Yang. 2008. “Flexible Latent Variable Models for Multi-Task Learning.” Machine Learning 73 (April). https://doi.org/10.1007/s10994-008-5050-1.
Zhang, Jingfeng, Xilie Xu, Bo Han, Gang Niu, Lizhen Cui, Masashi Sugiyama, and Mohan Kankanhalli. 2020. “Attacks Which Do Not Kill Training Make Adversarial Learning Stronger.” arXiv. https://doi.org/10.48550/ARXIV.2002.11242.
Zhang, Jing-Xuan, Zhen-Hua Ling, and Li-Rong Dai. 2020. “Recognition-Synthesis Based Non-Parallel Voice Conversion with Adversarial Learning.” Interspeech 2020, October. https://doi.org/10.21437/interspeech.2020-36.
Zhang, Jinjin, Wei Wang, Di Huang, Qingjie Liu, and Yunhong Wang. 2019. “A Feasible Framework for Arbitrary-Shaped Scene Text Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1912.04561.
Zhang, Junbo, Yu Zheng, and Dekang Qi. 2016. “Deep Spatio-Temporal Residual Networks for Citywide Crowd Flows Prediction.” arXiv. https://doi.org/10.48550/ARXIV.1610.00081.
Zhang, Junqi, Yiqun Liu, Shaoping Ma, and Qi Tian. 2018. “Relevance Estimation with Multiple Information Sources on Search Engine Result Pages.” Proceedings of the 27th ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3269206.3271673.
Zhang, Junzhe, Xinyi Chen, Zhongang Cai, Liang Pan, Haiyu Zhao, Shuai Yi, Chai Kiat Yeo, Bo Dai, and Chen Change Loy. 2021. “Unsupervised 3D Shape Completion Through GAN Inversion.” arXiv. https://doi.org/10.48550/ARXIV.2104.13366.
Zhang, Kai, Jingyun Liang, Luc Van Gool, and Radu Timofte. 2021. “Designing a Practical Degradation Model for Deep Blind Image Super-Resolution.” arXiv. https://doi.org/10.48550/ARXIV.2103.14006.
Zhang, Kaiqing, Zhuoran Yang, and Tamer Başar. 2019. “Multi-Agent Reinforcement Learning: A Selective Overview of Theories and Algorithms.” arXiv. https://doi.org/10.48550/ARXIV.1911.10635.
Zhang, Lingming, Yue Zhao, Deyu Meng, Zhiming Cui, Chenqiang Gao, Xinbo Gao, Chunfeng Lian, and Dinggang Shen. 2020. “TSGCNet: Discriminative Geometric Feature Learning with Two-Stream GraphConvolutional Network for 3D Dental Model Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.2012.13697.
Zhang, Linguang, Maciej Halber, and Szymon Rusinkiewicz. 2019. “Accelerating Large-Kernel Convolution Using Summed-Area Tables.” arXiv. https://doi.org/10.48550/ARXIV.1906.11367.
Zhang, Lin-Sen, Jian-Jun Tao, Guo-Hua Wang, and Xiao-Jing Zheng. 2021. “Experimental Study on the Origin of Lobe-Cleft Structures in a Sand Storm.” Acta Mechanica Sinica 37 (January). https://doi.org/10.1007/s10409-021-01053-7.
Zhang, Liqiang, Chengzhu Yu, Heng Lu, Chao Weng, Yusong Wu, Xiang Xie, Zijin Li, and Dong Yu. 2019. “Learning Singing from Speech.” arXiv. https://doi.org/10.48550/ARXIV.1912.10128.
Zhang, Liqiang, Chengzhu Yu, Heng Lu, Chao Weng, Chunlei Zhang, Yusong Wu, Xiang Xie, Zijin Li, and Dong Yu. 2020. “DurIAN-SC: Duration Informed Attention Network Based Singing Voice Conversion System.” arXiv. https://doi.org/10.48550/ARXIV.2008.03009.
Zhang, Liwen, John Winn, and Ryota Tomioka. 2016. “Gaussian Attention Model and Its Application to Knowledge Base Embedding and Question Answering.” arXiv. https://doi.org/10.48550/ARXIV.1611.02266.
Zhang, Ningyu, Luoqiu Li, Xiang Chen, Shumin Deng, Zhen Bi, Chuanqi Tan, Fei Huang, and Huajun Chen. 2021. “Differentiable Prompt Makes Pre-Trained Language Models Better Few-Shot Learners.” arXiv. https://doi.org/10.48550/ARXIV.2108.13161.
Zhang, Qiang, Zeyuan Wang, Yuqiang Han, Haoran Yu, Xurui Jin, and Huajun Chen. 2022. “Prompt-Guided Injection of Conformation to Pre-Trained Protein Model.” arXiv. https://doi.org/10.48550/ARXIV.2202.02944.
Zhang, Qingru, Minshuo Chen, Alexander Bukharin, Nikos Karampatziakis, Pengcheng He, Yu Cheng, Weizhu Chen, and Tuo Zhao. 2023. “AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2303.10512.
Zhang, Renrui, Ziyu Guo, Rongyao Fang, Bin Zhao, Dong Wang, Yu Qiao, Hongsheng Li, and Peng Gao. 2022. “Point-M2AE: Multi-Scale Masked Autoencoders for Hierarchical Point Cloud Pre-Training.” arXiv. https://doi.org/10.48550/ARXIV.2205.14401.
Zhang, Rui, and Joel Tetreault. 2019. “This Email Could Save Your Life: Introducing the Task of Email Subject Line Generation.” arXiv. https://doi.org/10.48550/ARXIV.1906.03497.
Zhang, Shichang, Yozen Liu, Neil Shah, and Yizhou Sun. 2022. “GStarX: Explaining Graph Neural Networks with Structure-Aware Cooperative Games.” arXiv. https://doi.org/10.48550/ARXIV.2201.12380.
Zhang, Shichang, Jiani Zhang, Xiang Song, Soji Adeshina, Da Zheng, Christos Faloutsos, and Yizhou Sun. 2023. “PaGE-Link: Path-Based Graph Neural Network Explanation for Heterogeneous Link Prediction.” Proceedings of the ACM Web Conference 2023, April. https://doi.org/10.1145/3543507.3583511.
Zhang, Shijie, and Gang Wu. 2021. “Efficient Online Log Parsing with Log Punctuations Signature.” Applied Sciences 11 (December). https://doi.org/10.3390/app112411974.
Zhang, Shijie, Hongzhi Yin, Tong Chen, Quoc Viet Nguyen Hung, Zi Huang, and Lizhen Cui. 2020. “GCN-Based User Representation Learning for Unifying Robust Recommendation and Fraudster Detection.” Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, July. https://doi.org/10.1145/3397271.3401165.
Zhang, Shun, Zhenfang Chen, Yikang Shen, Mingyu Ding, Joshua B. Tenenbaum, and Chuang Gan. 2023. “Planning with Large Language Models for Code Generation.” arXiv. https://doi.org/10.48550/ARXIV.2303.05510.
Zhang, Sixin, Anna Choromanska, and Yann LeCun. 2014. “Deep Learning with Elastic Averaging SGD.” arXiv. https://doi.org/10.48550/ARXIV.1412.6651.
Zhang, Tianyi, Faisal Ladhak, Esin Durmus, Percy Liang, Kathleen McKeown, and Tatsunori B. Hashimoto. 2023. “Benchmarking Large Language Models for News Summarization.” arXiv. https://doi.org/10.48550/ARXIV.2301.13848.
Zhang, Tianyun, Shaokai Ye, Kaiqi Zhang, Xiaolong Ma, Ning Liu, Linfeng Zhang, Jian Tang, et al. 2018. “StructADMM: A Systematic, High-Efficiency Framework of Structured Weight Pruning for DNNs.” arXiv. https://doi.org/10.48550/ARXIV.1807.11091.
Zhang, Tianyun, Shaokai Ye, Kaiqi Zhang, Jian Tang, Wujie Wen, Makan Fardad, and Yanzhi Wang. 2018. “A Systematic DNN Weight Pruning Framework Using Alternating Direction Method of Multipliers.” Computer Vision – ECCV 2018. https://doi.org/10.1007/978-3-030-01237-3_12.
Zhang, Tong. 2004. “Statistical Behavior and Consistency of Classification Methods Based on Convex Risk Minimization.” The Annals of Statistics 32 (February). https://doi.org/10.1214/aos/1079120130.
———. 2009. “Some Sharp Performance Bounds for Least Squares Regression with L1 Regularization.” The Annals of Statistics 37 (October). https://doi.org/10.1214/08-aos659.
Zhang, Weinan, Yifei Rong, Jun Wang, Tianchi Zhu, and Xiaofan Wang. 2016. “Feedback Control of Real-Time Display Advertising.” Proceedings of the Ninth ACM International Conference on Web Search and Data Mining, February. https://doi.org/10.1145/2835776.2835843.
Zhang, Weinan, and Jun Wang. 2015. “Statistical Arbitrage Mining for Display Advertising.” Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2783258.2783269.
Zhang, Weinan, Shuai Yuan, and Jun Wang. 2014. “Optimal Real-Time Bidding for Display Advertising.” Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August. https://doi.org/10.1145/2623330.2623633.
Zhang, Weinan, Shuai Yuan, Jun Wang, and Xuehua Shen. 2014. “Real-Time Bidding Benchmarking with iPinYou Dataset.” arXiv. https://doi.org/10.48550/ARXIV.1407.7073.
Zhang, Wen, Bibek Paudel, Wei Zhang, Abraham Bernstein, and Huajun Chen. 2019. “Interaction Embeddings for Prediction and Explanation in Knowledge Graphs.” Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, January. https://doi.org/10.1145/3289600.3291014.
Zhang, Xiang, and Yann LeCun. 2015. “Text Understanding from Scratch.” arXiv. https://doi.org/10.48550/ARXIV.1502.01710.
Zhang, Xiang, Junbo Zhao, and Yann LeCun. 2015. “Character-Level Convolutional Networks for Text Classification.” arXiv. https://doi.org/10.48550/ARXIV.1509.01626.
Zhang, Xinlu, Yujie Lu, Weizhi Wang, An Yan, Jun Yan, Lianke Qin, Heng Wang, Xifeng Yan, William Yang Wang, and Linda Ruth Petzold. 2023. “GPT-4V(ision) as a Generalist Evaluator for Vision-Language Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2311.01361.
Zhang, Xintong, Shasha Li, Bin Ji, and Ting Wang. 2023. “Learning Well-Separated and Representative Prototypes for Few-Shot Event Detection.” Natural Language Processing and Chinese Computing. https://doi.org/10.1007/978-3-031-44696-2_23.
Zhang, Xu, Yong Xu, Qingwei Lin, Bo Qiao, Hongyu Zhang, Yingnong Dang, Chunyu Xie, et al. 2019. “Robust Log-Based Anomaly Detection on Unstable Log Data.” Proceedings of the 2019 27th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, August. https://doi.org/10.1145/3338906.3338931.
Zhang, Yabin, Hui Tang, Kui Jia, and Mingkui Tan. 2019. “Domain-Symmetric Networks for Adversarial Domain Adaptation.” arXiv. https://doi.org/10.48550/ARXIV.1904.04663.
Zhang, Ya-Jie, Shifeng Pan, Lei He, and Zhen-Hua Ling. 2018. “Learning Latent Representations for Style Control and Transfer in End-to-End Speech Synthesis.” arXiv. https://doi.org/10.48550/ARXIV.1812.04342.
Zhang, Ya, Yi Wei, and Jianbiao Ren. 2014. “Multi-Touch Attribution in Online Advertising with Survival Theory.” 2014 IEEE International Conference on Data Mining, December. https://doi.org/10.1109/icdm.2014.130.
Zhang, Yi-Fan, Hanlin Zhang, Zachary C. Lipton, Li Erran Li, and Eric P. Xing. 2022. “Exploring Transformer Backbones for Heterogeneous Treatment Effect Estimation.” arXiv. https://doi.org/10.48550/ARXIV.2202.01336.
Zhang, Yi, Jianguo Lu, and Ofer Shai. 2018. “Improve Network Embeddings with Regularization.” Proceedings of the 27th ACM International Conference on Information and Knowledge Management, October. https://doi.org/10.1145/3269206.3269320.
Zhang, Yiming, Yujie Fan, Wei Song, Shifu Hou, Yanfang Ye, Xin Li, Liang Zhao, Chuan Shi, Jiabin Wang, and Qi Xiong. 2019. “Your Style Your Identity: Leveraging Writing and Photography Styles for Drug Trafficker Identification in Darknet Markets over Attributed Heterogeneous Information Network.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313537.
Zhang, Yiming, Yujie Fan, Yanfang Ye, Liang Zhao, and Chuan Shi. 2019. “Key Player Identification in Underground Forums over Attributed Heterogeneous Information Network Embedding Framework.” Proceedings of the 28th ACM International Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3357384.3357876.
Zhang, Ying, Tao Xiang, Timothy M. Hospedales, and Huchuan Lu. 2017. “Deep Mutual Learning.” arXiv. https://doi.org/10.48550/ARXIV.1706.00384.
Zhang, Yingxue, Soumyasundar Pal, Mark Coates, and Deniz Üstebay. 2018. “Bayesian Graph Convolutional Neural Networks for Semi-Supervised Classification.” arXiv. https://doi.org/10.48550/ARXIV.1811.11103.
Zhang, Yi, Dapeng Zhang, and Haoyu Jiang. 2023. “A Review of Offshore Wind and Wave Installations in Some Areas with an Eye Towards Generating Economic Benefits and Offering Commercial Inspiration.” Sustainability 15 (May). https://doi.org/10.3390/su15108429.
Zhang, Yuan-Hang, and Massimiliano Di Ventra. 2023. “Transformer Quantum State: A Multipurpose Model for Quantum Many-Body Problems.” Physical Review B 107 (February). https://doi.org/10.1103/physrevb.107.075147.
Zhang, Yuan, Dong Wang, and Yan Zhang. 2019. “Neural IR Meets Graph Embedding: A Ranking Model for Product Search.” arXiv. https://doi.org/10.48550/ARXIV.1901.08286.
Zhang, Yuanzhe, Kang Liu, Shizhu He, Guoliang Ji, Zhanyi Liu, Hua Wu, and Jun Zhao. 2016. “Question Answering over Knowledge Base with Neural Attention Combining Global Knowledge Information.” arXiv. https://doi.org/10.48550/ARXIV.1606.00979.
Zhang, Yu, William Chan, and Navdeep Jaitly. 2016. “Very Deep Convolutional Networks for End-to-End Speech Recognition.” arXiv. https://doi.org/10.48550/ARXIV.1610.03022.
Zhang, Yuekai, Sining Sun, and Long Ma. 2021. “Tiny Transducer: A Highly-Efficient Speech Recognition Model on Edge Devices.” arXiv. https://doi.org/10.48550/ARXIV.2101.06856.
Zhang, Yu, Wei Han, James Qin, Yongqiang Wang, Ankur Bapna, Zhehuai Chen, Nanxin Chen, et al. 2023. “Google USM: Scaling Automatic Speech Recognition Beyond 100 Languages.” arXiv. https://doi.org/10.48550/ARXIV.2303.01037.
Zhang, Yuhao, Peng Qi, and Christopher D. Manning. 2018. “Graph Convolution over Pruned Dependency Trees Improves Relation Extraction.” arXiv. https://doi.org/10.48550/ARXIV.1809.10185.
Zhang, Yunpeng, Wenzhao Zheng, Zheng Zhu, Guan Huang, Jie Zhou, and Jiwen Lu. 2022. “A Simple Baseline for Multi-Camera 3D Object Detection.” arXiv. https://doi.org/10.48550/ARXIV.2208.10035.
Zhang, Zewang, Yibin Zheng, Xinhui Li, and Li Lu. 2022. “WeSinger 2: Fully Parallel Singing Voice Synthesis via Multi-Singer Conditional Adversarial Training.” arXiv. https://doi.org/10.48550/ARXIV.2207.01886.
Zhang, Zhenyu, Bowen Yu, Xiaobo Shu, Tingwen Liu, Hengzhu Tang, Wang Yubin, and Li Guo. 2020. “Document-Level Relation Extraction with Dual-Tier Heterogeneous Graph.” Proceedings of the 28th International Conference on Computational Linguistics. https://doi.org/10.18653/v1/2020.coling-main.143.
Zhang, Zhihao, Alan Zhu, Lijie Yang, Yihua Xu, Lanting Li, Phitchaya Mangpo Phothilimthana, and Zhihao Jia. 2024. “Accelerating Retrieval-Augmented Language Model Serving with Speculation.” arXiv. https://doi.org/10.48550/ARXIV.2401.14021.
Zhang, Zhiyuan, Binh-Son Hua, David W. Rosen, and Sai-Kit Yeung. 2019. “Rotation Invariant Convolutions for 3D Point Clouds Deep Learning.” arXiv. https://doi.org/10.48550/ARXIV.1908.06297.
Zhang, Zi-Ke, Tao Zhou, and Yi-Cheng Zhang. 2011. “Tag-Aware Recommender Systems: A State-of-the-Art Survey.” Journal of Computer Science and Technology 26 (September). https://doi.org/10.1007/s11390-011-0176-1.
Zhang, Ziqiang, Long Zhou, Chengyi Wang, Sanyuan Chen, Yu Wu, Shujie Liu, Zhuo Chen, et al. 2023. “Speak Foreign Languages with Your Own Voice: Cross-Lingual Neural Codec Language Modeling.” arXiv. https://doi.org/10.48550/ARXIV.2303.03926.
Zhao, Bo, Benjamin I. P. Rubinstein, Jim Gemmell, and Jiawei Han. 2012. “A Bayesian Approach to Discovering Truth from Conflicting Sources for Data Integration.” arXiv. https://doi.org/10.48550/ARXIV.1203.0058.
Zhao, Bowen, Hannaneh Hajishirzi, and Qingqing Cao. 2024. “APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inference.” arXiv. https://doi.org/10.48550/ARXIV.2401.12200.
Zhao, Huasha, Luo Si, Xiaogang Li, and Qiong Zhang. 2017. “Recommending Complementary Products in e-Commerce Push Notifications with a Mixture Model Approach.” arXiv. https://doi.org/10.48550/ARXIV.1707.08113.
Zhao, Jinming, Ruichen Li, Qin Jin, Xinchao Wang, and Haizhou Li. 2021. “MEmoBERT: Pre-Training Model with Prompt-Based Learning for Multimodal Emotion Recognition.” arXiv. https://doi.org/10.48550/ARXIV.2111.00865.
Zhao, Junbo, Michael Mathieu, and Yann LeCun. 2016. “Energy-Based Generative Adversarial Network.” arXiv. https://doi.org/10.48550/ARXIV.1609.03126.
Zhao, Jun, Guang Qiu, Ziyu Guan, Wei Zhao, and Xiaofei He. 2018. “Deep Reinforcement Learning for Sponsored Search Real-Time Bidding.” arXiv. https://doi.org/10.48550/ARXIV.1803.00259.
Zhao, Lingxiao, Saurabh Sawlani, Arvind Srinivasan, and Leman Akoglu. 2022. “Graph Anomaly Detection with Unsupervised GNNs.” arXiv. https://doi.org/10.48550/ARXIV.2210.09535.
Zhao, Nengwen, Jing Zhu, Rong Liu, Dapeng Liu, Ming Zhang, and Dan Pei. 2019. “Label-Less: A Semi-Automatic Labelling Tool for KPI Anomalies.” IEEE INFOCOM 2019 - IEEE Conference on Computer Communications, April. https://doi.org/10.1109/infocom.2019.8737429.
Zhao, Tiancheng, and Maxine Eskenazi. 2016. “Towards End-to-End Learning for Dialog State Tracking and Management Using Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1606.02560.
Zhao, Tong, Bo Ni, Wenhao Yu, Zhichun Guo, Neil Shah, and Meng Jiang. 2021. “Action Sequence Augmentation for Early Graph-Based Anomaly Detection.” Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management, October. https://doi.org/10.1145/3459637.3482313.
Zhao, Wayne Xin, Kun Zhou, Junyi Li, Tianyi Tang, Xiaolei Wang, Yupeng Hou, Yingqian Min, et al. 2023. “A Survey of Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2303.18223.
Zhao, Weixiang, Shilong Wang, Yulin Hu, Yanyan Zhao, Bing Qin, Xuanyu Zhang, Qing Yang, Dongliang Xu, and Wanxiang Che. 2024. “DAPT: A Dual Attention Framework for Parameter-Efficient Continual Learning of Large Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2401.08295.
Zhao, Wenliang, Yongming Rao, Zuyan Liu, Benlin Liu, Jie Zhou, and Jiwen Lu. 2023. “Unleashing Text-to-Image Diffusion Models for Visual Perception.” arXiv. https://doi.org/10.48550/ARXIV.2303.02153.
Zhao, Xuandong, Yu-Xiang Wang, and Lei Li. 2023. “Protecting Language Generation Models via Invisible Watermarking.” arXiv. https://doi.org/10.48550/ARXIV.2302.03162.
Zhao, Ying, and George Karypis. 2002. “Evaluation of Hierarchical Clustering Algorithms for Document Datasets.” Proceedings of the Eleventh International Conference on Information and Knowledge Management, November. https://doi.org/10.1145/584792.584877.
Zhao, Yiren, Xitong Gao, Daniel Bates, Robert Mullins, and Cheng-Zhong Xu. 2019. “Focused Quantization for Sparse CNNs.” arXiv. https://doi.org/10.48550/ARXIV.1903.03046.
Zhao, Y., J. Staudenmayer, B. A. Coull, and M. P. Wand. 2006. “General Design Bayesian Generalized Linear Mixed Models.” Statistical Science 21 (February). https://doi.org/10.1214/088342306000000015.
Zhao, Yue, Long Zhao, Xingyi Zhou, Jialin Wu, Chun-Te Chu, Hui Miao, Florian Schroff, et al. 2024. “Distilling Vision-Language Models on Millions of Videos.” arXiv. https://doi.org/10.48550/ARXIV.2401.06129.
Zhao, Zheng, and Huan Liu. 2007. “Semi-Supervised Feature Selection via Spectral Analysis.” Proceedings of the 2007 SIAM International Conference on Data Mining, April. https://doi.org/10.1137/1.9781611972771.75.
Zharov, Yaroslav, Denis Korzhenkov, Pavel Shvechikov, and Alexander Tuzhilin. 2018. “YASENN: Explaining Neural Networks via Partitioning Activation Sequences.” arXiv. https://doi.org/10.48550/ARXIV.1811.02783.
Zheng, Jianming, Fei Cai, Yanxiang Ling, and Honghui Chen. 2020. “Heterogeneous Graph Neural Networks to Predict What Happen Next.” Proceedings of the 28th International Conference on Computational Linguistics. https://doi.org/10.18653/v1/2020.coling-main.29.
Zheng, Junwen, and Martin Fischer. 2023. “BIM-GPT: A Prompt-Based Virtual Assistant Framework for BIM Information Retrieval.” arXiv. https://doi.org/10.48550/ARXIV.2304.09333.
Zheng, Liang, Yi Yang, and Alexander G. Hauptmann. 2016. “Person Re-Identification: Past, Present and Future.” arXiv. https://doi.org/10.48550/ARXIV.1610.02984.
Zheng, Ou, Mohamed Abdel-Aty, Dongdong Wang, Zijin Wang, and Shengxuan Ding. 2023. “ChatGPT Is on the Horizon: Could a Large Language Model Be Suitable for Intelligent Traffic Safety Research and Applications?” arXiv. https://doi.org/10.48550/ARXIV.2303.05382.
Zheng, Qinkai, Xiao Xia, Xu Zou, Yuxiao Dong, Shan Wang, Yufei Xue, Zihan Wang, et al. 2023. “CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Evaluations on HumanEval-x.” arXiv. https://doi.org/10.48550/ARXIV.2303.17568.
Zheng, Shun, Xu Han, Yankai Lin, Peilin Yu, Lu Chen, Ling Huang, Zhiyuan Liu, and Wei Xu. 2018. “DIAG-NRE: A Neural Pattern Diagnosis Framework for Distantly Supervised Neural Relation Extraction.” arXiv. https://doi.org/10.48550/ARXIV.1811.02166.
Zheng, Yu, Licia Capra, Ouri Wolfson, and Hai Yang. 2014. “Urban Computing.” ACM Transactions on Intelligent Systems and Technology 5 (September). https://doi.org/10.1145/2629592.
Zheng, Yu, Chen Gao, Liang Chen, Depeng Jin, and Yong Li. 2021. “DGCN: Diversified Recommendation with Graph Convolutional Networks.” Proceedings of the Web Conference 2021, April. https://doi.org/10.1145/3442381.3449835.
Zheng, Zhaoheng, Jingmin Wei, Xuefeng Hu, Haidong Zhu, and Ram Nevatia. 2023. “Large Language Models Are Good Prompt Learners for Low-Shot Image Classification.” arXiv. https://doi.org/10.48550/ARXIV.2312.04076.
Zheng, Zibin, Weili Chen, Zhijie Zhong, Zhiguang Chen, and Yutong Lu. 2023. “Securing the Ethereum from Smart Ponzi Schemes: Identification Using Static Features.” ACM Transactions on Software Engineering and Methodology 32 (July). https://doi.org/10.1145/3571847.
Zheng, Zijian, and Geoffrey I. Webb. 2000. Machine Learning 41. https://doi.org/10.1023/a:1007613203719.
Zhiheng, Kang, and Li Ning. 2019. “PyramNet: Point Cloud Pyramid Attention Network and Graph Embedding Module for Classification and Segmentation.” arXiv. https://doi.org/10.48550/ARXIV.1906.03299.
Zhou, Aojun, Anbang Yao, Kuan Wang, and Yurong Chen. 2018. “Explicit Loss-Error-Aware Quantization for Low-Bit Deep Neural Networks.” 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, June. https://doi.org/10.1109/cvpr.2018.00982.
Zhou, Bowen, and Shahriar Shariat. 2016. “Finding Needle in a Million Metrics: Anomaly Detection in a Large-Scale Computational Advertising Platform.” arXiv. https://doi.org/10.48550/ARXIV.1602.07057.
Zhou, Ce, Qian Li, Chen Li, Jun Yu, Yixin Liu, Guangjing Wang, Kai Zhang, et al. 2023. “A Comprehensive Survey on Pretrained Foundation Models: A History from BERT to ChatGPT.” arXiv. https://doi.org/10.48550/ARXIV.2302.09419.
Zhou, Chunting, Pengfei Liu, Puxin Xu, Srini Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, et al. 2023. “LIMA: Less Is More for Alignment.” arXiv. https://doi.org/10.48550/ARXIV.2305.11206.
Zhou, Daquan, Xiaojie Jin, Qibin Hou, Kaixin Wang, Jianchao Yang, and Jiashi Feng. 2019. “Neural Epitome Search for Architecture-Agnostic Network Compression.” arXiv. https://doi.org/10.48550/ARXIV.1907.05642.
Zhou, Ding-Xuan. 2018. “Universality of Deep Convolutional Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.1805.10769.
Zhou, Han, Xingchen Wan, Lev Proleev, Diana Mincu, Jilin Chen, Katherine Heller, and Subhrajit Roy. 2023. “Batch Calibration: Rethinking Calibration for in-Context Learning and Prompt Engineering.” arXiv. https://doi.org/10.48550/ARXIV.2309.17249.
Zhou, Hao, Tom Young, Minlie Huang, Haizhou Zhao, Jingfang Xu, and Xiaoyan Zhu. 2018. “Commonsense Knowledge Aware Conversation Generation with Graph Attention.” Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July. https://doi.org/10.24963/ijcai.2018/643.
Zhou, Hong-Cai, Jeffrey R. Long, and Omar M. Yaghi. 2012. “Introduction to Metal–Organic Frameworks.” Chemical Reviews 112 (January). https://doi.org/10.1021/cr300014x.
Zhou, Jiawei, Yixuan Zhang, Qianni Luo, Andrea G Parker, and Munmun De Choudhury. 2023. “Synthetic Lies: Understanding AI-Generated Misinformation and Evaluating Algorithmic and Human Solutions.” Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, April. https://doi.org/10.1145/3544548.3581318.
Zhou, Jinghao, Chen Wei, Huiyu Wang, Wei Shen, Cihang Xie, Alan Yuille, and Tao Kong. 2021. “iBOT: Image BERT Pre-Training with Online Tokenizer.” arXiv. https://doi.org/10.48550/ARXIV.2111.07832.
Zhou, Kai, and Yevgeniy Vorobeychik. 2020. “Robust Collective Classification Against Structural Attacks.” arXiv. https://doi.org/10.48550/ARXIV.2007.13073.
Zhou, Ming, Ziyu Wan, Hanjing Wang, Muning Wen, Runzhe Wu, Ying Wen, Yaodong Yang, Weinan Zhang, and Jun Wang. 2021. “MALib: A Parallel Framework for Population-Based Multi-Agent Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.2106.07551.
Zhou, Rensheng R., Nicoleta Serban, and Nagi Gebraeel. 2011. “Degradation Modeling Applied to Residual Lifetime Prediction Using Functional Data Analysis.” The Annals of Applied Statistics 5 (June). https://doi.org/10.1214/10-aoas448.
Zhou, Ruida, Tao Liu, Min Cheng, Dileep Kalathil, P. R. Kumar, and Chao Tian. 2023. “Natural Actor-Critic for Robust Reinforcement Learning with Function Approximation.” arXiv. https://doi.org/10.48550/ARXIV.2307.08875.
Zhou, Shuchang, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, and Yuheng Zou. 2016. “DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients.” arXiv. https://doi.org/10.48550/ARXIV.1606.06160.
Zhou, Xuanyu, Charles R. Qi, Yin Zhou, and Dragomir Anguelov. 2022. “RIDDLE: Lidar Data Compression with Range Image Deep Delta Encoding.” arXiv. https://doi.org/10.48550/ARXIV.2206.01738.
Zhou, Yilun, Steven Schockaert, and Julie Shah. 2019. “Predicting ConceptNet Path Quality Using Crowdsourced Assessments of Naturalness.” The World Wide Web Conference, May. https://doi.org/10.1145/3308558.3313486.
Zhou, Yiyang, Chenhang Cui, Jaehong Yoon, Linjun Zhang, Zhun Deng, Chelsea Finn, Mohit Bansal, and Huaxiu Yao. 2023. “Analyzing and Mitigating Object Hallucination in Large Vision-Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2310.00754.
Zhou, Yuefu, Ya Zhang, Yanfeng Wang, and Qi Tian. 2018. “Accelerate CNN via Recursive Bayesian Pruning.” arXiv. https://doi.org/10.48550/ARXIV.1812.00353.
Zhou, Zhi-Hua, Yu-Yin Sun, and Yu-Feng Li. 2008. “Multi-Instance Learning by Treating Instances as Non-i.i.d. Samples.” arXiv. https://doi.org/10.48550/ARXIV.0807.1997.
Zhu, Benjin, Junqiang Huang, Zeming Li, Xiangyu Zhang, and Jian Sun. 2020. “EqCo: Equivalent Rules for Self-Supervised Contrastive Learning.” arXiv. https://doi.org/10.48550/ARXIV.2010.01929.
Zhu, Chenzhuo, Song Han, Huizi Mao, and William J. Dally. 2016. “Trained Ternary Quantization.” arXiv. https://doi.org/10.48550/ARXIV.1612.01064.
Zhu, Dingyuan, Ziwei Zhang, Peng Cui, and Wenwu Zhu. 2019. “Robust Graph Convolutional Networks Against Adversarial Attacks.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330851.
Zhu, Fei, Fei Ye, Yuchen Fu, Quan Liu, and Bairong Shen. 2019. “Electrocardiogram Generation with a Bidirectional LSTM-CNN Generative Adversarial Network.” Scientific Reports 9 (May). https://doi.org/10.1038/s41598-019-42516-z.
Zhu, Guangxu, Dongzhu Liu, Yuqing Du, Changsheng You, Jun Zhang, and Kaibin Huang. 2018. “Towards an Intelligent Edge: Wireless Communication Meets Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1809.00343.
Zhu, Jun-Jie, Jinyue Jiang, Meiqi Yang, and Zhiyong Jason Ren. 2023. “ChatGPT and Environmental Research.” Environmental Science &Amp; Technology 57 (March). https://doi.org/10.1021/acs.est.3c01818.
Zhu, Lei, Xinjiang Wang, Zhanghan Ke, Wayne Zhang, and Rynson Lau. 2023. “BiFormer: Vision Transformer with Bi-Level Routing Attention.” arXiv. https://doi.org/10.48550/ARXIV.2303.08810.
Zhu, Linhong, Dong Guo, Junming Yin, Greg Ver Steeg, and Aram Galstyan. 2016. “Scalable Temporal Latent Space Inference for Link Prediction in Dynamic Social Networks.” IEEE Transactions on Knowledge and Data Engineering 28 (October). https://doi.org/10.1109/tkde.2016.2591009.
Zhu, Qiannan, Xiaofei Zhou, Jia Wu, Jianlong Tan, and Li Guo. 2019. “Neighborhood-Aware Attentional Representation for Multilingual Knowledge Graphs.” Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, August. https://doi.org/10.24963/ijcai.2019/269.
Zhu, Qihao, and Jianxi Luo. 2022. “Generative Transformers for Design Concept Generation.” arXiv. https://doi.org/10.48550/ARXIV.2211.03468.
Zhu, Qihao, Xinyu Zhang, and Jianxi Luo. 2022. “Biologically Inspired Design Concept Generation Using Generative Pre-Trained Transformers.” arXiv. https://doi.org/10.48550/ARXIV.2212.13196.
Zhu, Rong, Kun Zhao, Hongxia Yang, Wei Lin, Chang Zhou, Baole Ai, Yong Li, and Jingren Zhou. 2019. “AliGraph: A Comprehensive Graph Neural Network Platform.” arXiv. https://doi.org/10.48550/ARXIV.1902.08730.
Zhu, Rui-Jie, Qihang Zhao, Guoqi Li, and Jason K. Eshraghian. 2023. “SpikeGPT: Generative Pre-Trained Language Model with Spiking Neural Networks.” arXiv. https://doi.org/10.48550/ARXIV.2302.13939.
Zhu, Xiaodan, Parinaz Sobhani, and Hongyu Guo. 2015. “Long Short-Term Memory over Tree Structures.” arXiv. https://doi.org/10.48550/ARXIV.1503.04881.
Zhu, Yao, Hongzhi Liu, Zhonghai Wu, and Yingpeng Du. 2020. “Relation-Aware Neighborhood Matching Model for Entity Alignment.” arXiv. https://doi.org/10.48550/ARXIV.2012.08128.
Zhu, Yiming, Peixian Zhang, Ehsan-Ul Haq, Pan Hui, and Gareth Tyson. 2023. “Can ChatGPT Reproduce Human-Generated Labels? A Study of Social Computing Tasks.” arXiv. https://doi.org/10.48550/ARXIV.2304.10145.
Zhu, Yi, Xinke Zhou, Jipeng Qiang, Yun Li, Yunhao Yuan, and Xindong Wu. 2022. “Prompt-Learning for Short Text Classification.” arXiv. https://doi.org/10.48550/ARXIV.2202.11345.
Zhu, Yuke, Roozbeh Mottaghi, Eric Kolve, Joseph J. Lim, Abhinav Gupta, Li Fei-Fei, and Ali Farhadi. 2016. “Target-Driven Visual Navigation in Indoor Scenes Using Deep Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1609.05143.
Zhu, Yun, Jianhao Guo, and Siliang Tang. 2023. “SGL-PT: A Strong Graph Learner with Graph Prompt Tuning.” arXiv. https://doi.org/10.48550/ARXIV.2302.12449.
Zhu, Yutao, Huaying Yuan, Shuting Wang, Jiongnan Liu, Wenhan Liu, Chenlong Deng, Haonan Chen, Zhicheng Dou, and Ji-Rong Wen. 2023. “Large Language Models for Information Retrieval: A Survey.” arXiv. https://doi.org/10.48550/ARXIV.2308.07107.
Zhu, Zijian, Yichi Zhang, Hai Chen, Yinpeng Dong, Shu Zhao, Wenbo Ding, Jiachen Zhong, and Shibao Zheng. 2023. “Understanding the Robustness of 3D Object Detection with Bird’s-Eye-View Representations in Autonomous Driving.” arXiv. https://doi.org/10.48550/ARXIV.2303.17297.
Zhuang, Chengxu, Alex Lin Zhai, and Daniel Yamins. 2019. “Local Aggregation for Unsupervised Learning of Visual Embeddings.” arXiv. https://doi.org/10.48550/ARXIV.1903.12355.
Zhuang, Tao, Wenwu Ou, and Zhirong Wang. 2018. “Globally Optimized Mutual Influence Aware Ranking in e-Commerce Search.” arXiv. https://doi.org/10.48550/ARXIV.1805.08524.
Zhuang, Yan, Guoliang Li, Zhuojian Zhong, and Jianhua Feng. 2017. “Hike.” Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, November. https://doi.org/10.1145/3132847.3132912.
Zimmermann, Roland S., Yash Sharma, Steffen Schneider, Matthias Bethge, and Wieland Brendel. 2021. “Contrastive Learning Inverts the Data Generating Process.” arXiv. https://doi.org/10.48550/ARXIV.2102.08850.
Zliobaite, Indre. 2015. “A Survey on Measuring Indirect Discrimination in Machine Learning.” arXiv. https://doi.org/10.48550/ARXIV.1511.00148.
Zoph, Barret, Golnaz Ghiasi, Tsung-Yi Lin, Yin Cui, Hanxiao Liu, Ekin D. Cubuk, and Quoc V. Le. 2020. “Rethinking Pre-Training and Self-Training.” arXiv. https://doi.org/10.48550/ARXIV.2006.06882.
Zoph, Barret, and Quoc V. Le. 2016. “Neural Architecture Search with Reinforcement Learning.” arXiv. https://doi.org/10.48550/ARXIV.1611.01578.
Zoph, Barret, Deniz Yuret, Jonathan May, and Kevin Knight. 2016. “Transfer Learning for Low-Resource Neural Machine Translation.” arXiv. https://doi.org/10.48550/ARXIV.1604.02201.
Zou, Andy, Zifan Wang, Nicholas Carlini, Milad Nasr, J. Zico Kolter, and Matt Fredrikson. 2023. “Universal and Transferable Adversarial Attacks on Aligned Language Models.” arXiv. https://doi.org/10.48550/ARXIV.2307.15043.
Zou, Lixin, Shengqiang Zhang, Hengyi Cai, Dehong Ma, Suqi Cheng, Daiting Shi, Zhifan Zhu, et al. 2021. “Pre-Trained Language Model Based Ranking in Baidu Search.” arXiv. https://doi.org/10.48550/ARXIV.2105.11108.
Zügner, Daniel, Amir Akbarnejad, and Stephan Günnemann. 2018. “Adversarial Attacks on Neural Networks for Graph Data.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220078.
Zügner, Daniel, and Stephan Günnemann. 2019. “Certifiable Robustness and Robust Training for Graph Convolutional Networks.” Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3292500.3330905.
Zuluaga-Gomez, Juan, Amrutha Prasad, Iuliia Nigmatulina, Saeed Sarfjoo, Petr Motlicek, Matthias Kleinert, Hartmut Helmke, Oliver Ohneiser, and Qingran Zhan. 2022. “How Does Pre-Trained Wav2Vec 2.0 Perform on Domain Shifted ASR? An Extensive Benchmark on Air Traffic Control Communications.” arXiv. https://doi.org/10.48550/ARXIV.2203.16822.
Zuo, Yuan, Guannan Liu, Hao Lin, Jia Guo, Xiaoqian Hu, and Junjie Wu. 2018. “Embedding Temporal Network via Neighborhood Formation.” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining, July. https://doi.org/10.1145/3219819.3220054.
Zwanenburg, Alex, Martin Vallières, Mahmoud A. Abdalah, Hugo J. W. L. Aerts, Vincent Andrearczyk, Aditya Apte, Saeed Ashrafinia, et al. 2020. “The Image Biomarker Standardization Initiative: Standardized Quantitative Radiomics for High-Throughput Image-Based Phenotyping.” Radiology 295 (May). https://doi.org/10.1148/radiol.2020191145.
Zweig, Geoffrey, Chengzhu Yu, Jasha Droppo, and Andreas Stolcke. 2017. “Advances in All-Neural Speech Recognition.” 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), March. https://doi.org/10.1109/icassp.2017.7953069.
Zwerts, Joeri A., Jelle Treep, Casper S. Kaandorp, Floor Meewis, Amparo C. Koot, and Heysem Kaya. 2021. “Introducing a Central African Primate Vocalisation Dataset for Automated Species Classification.” arXiv. https://doi.org/10.48550/ARXIV.2101.10390.